Google Cloud Container Registry is a place to manage Docker images, perform vulnerability analysis, and decide who can access what with fine-grained access control. Existing CI/CD integrations let users set up fully automated Docker pipelines.
N/A
LXD
Score 9.9 out of 10
N/A
LXD is a system container and virtual machine manager. It offers a unified user experience around full Linux systems running inside containers or virtual machines. LXD is image based and provides images for a wide number of Linux distributions. It supports various use cases, with support for different storage backends and network types and the option to install on hardware ranging from an individual laptop or cloud instance to a full server rack. LXD is written in Go. It is free software…
N/A
Red Hat OpenShift
Score 9.2 out of 10
N/A
OpenShift is Red Hat's Cloud Computing Platform as a Service (PaaS) offering. OpenShift is an application platform in the cloud where application developers and teams can build, test, deploy, and run their applications.
As a Cloud Engineer while working on a migration project we used GCR and my experience of using it is actually good the reason behind this is: 1: GCR uses industry-standard encryption to protect your data.2: GCR offers data loss prevention features to help you prevent sensitive data from being leaked or exposed and last but not the least is GCR provides audit logging so you can track who has accessed your data and when, because of these reasons its my go to tool.
Linux Containers LXD works very well when you, as a software developer using Linux, need to spin up a development environment to build and work on a specific piece of software. For example, you need to build some software that is tested and developed target Ubuntu 18.04. You're able to quickly create a container for Ubuntu 18.04 and log into it to run commands and build your software. This is easier to do than setting a virtual machine - e.g - via Virtual Box, but is a lot clunkier than doing it via Docker containers which give you a much more flexible configuration of the environment and are easier to start, stop, connect and share with other developers.
Red Hat OpenShift, despite its complexity and overhead, remains the most complete and enterprise-ready Kubernetes platform available. It excels in research projects like ours, where we need robust CI/CD, GPU scheduling, and tight integration with tools like Jupyter, OpenDataHub, and Quiskit. Its security, scalability, and operator ecosystem make it ideal for experimental and production-grade AI workloads. However, for simpler general hosting tasks—such as serving static websites or lightweight backend services—we find traditional VMs, Docker, or LXD more practical and resource-efficient. Red Hat OpenShift shines in complex, container-native workflows, but can be overkill for basic infrastructure needs.
We had a few microservices that dealt with notifications and alerts. We used OpenShift to deploy these microservices, which handle and deliver notifications using publish-subscribe models.
We had to expose an API to consumers via MTLS, which was implemented using Server secret integration in OpenShift. We were then able to deploy the APIs on OpenShift with API security.
We integrated Splunk with OpenShift to view the logs of our applications and gain real-time insights into usage, as well as provide high availability.
I wouldn't necessarily say there is look everyday technology transform. I can see a trend wherein Red Hat OpenShift is adopting all the new technology trends and helping their customers align with their priorities and the emerging technology trends. I wouldn't call out various scope for development every day. There is scope for development. It is all how the organizations adopt it and how they deliver it to their customers. I don't want to call out there is scope for development. It's happening. It is a never ending process.
At the moment, I don't have anything to call out. We are experiencing Red Hat OpenShift and we can see every day they're coming up with new features as and when they come up with new features, we want to experience it more and more. We are looking for opportunities wherein this can be leveraged to help our users and partners.
This is the current strategy for the company, most of the products in the organisation are aligning to Openshift and various use cases it support. Also lot of applications are being developed for AI use case, openshift.AI provides opportunity to host and leverage the AI capabilities for these applications
As I said before, the obserability is one of the weakest point of OpenShift and that has a lot to do with usability. The Kibana console is not fully integrated with OpenShift console and you have to switch from tab to tab to use it. Same with Prometheus, Jaeger and Grafan, it's a "simple" integration but if you want to do complex queries or dashboards you have to go to the specific console
Redhat openshift is generally reliable and available platform, it ensures high availability for most the situations. in fact the product where we put openshift in a box, we ensure that the availability is also happening at node and network level and also at storage level, so some of the factors that are outside of Openshift realm are also working in HA manner.
Overall, this platform is beneficial. The only downsides we have encountered have been with pods that occasionally hang. This results in resources being dedicated to dead or zombie pods. Over time, these wasted resources occasionally cause us issues, and we have had difficulty monitoring these pods. However, this issue does not overshadow the benefits we get from Openshift.
Their customer support team is good and quick to respond. On a couple of occassions, they have helped us in solving some issues which we were finding a tad difficult to comprehend. On a rare occasion, the response was a bit slow but maybe it was because of the festival season. Overall a good experience on this front.
I was not involved in the in person training, so i can not answer this question, but the team in my org worked directly with Openshift and able to get the in person training done easily, i did not hear problem or complain in this space, so i hope things happen seamlessly without any issue.
We went thru the training material on RH webesite, i think its very descriptive and the handson lab sesssions are very useful. It would be good to create more short duration videos covering one single aspect of openshift, this wll keep the interest and also it breaks down the complexity to reasonable chunks.
It is very easy to integrate GCR with other services and I integrated GCR with GKE and Cloud Build. 1: While working on one project I created one pipeline pulls the app's Docker image from GCR and builds the app's Docker image. Deployed that apps image in GKE. 2: Ive stored the data in GCR and that data was being used by cloud Run applications
Linux Containers LXD feels like a more primitive version of docker, docker-compose and similar projects from the docker ecosystem. The Dockerfile and docker-compose.yml methods of specifying a container setup, as well as the network and file configurations afforded by Docker make working with containers much easier and more reproducible than with Linux Containers LXD.
The Tanzu Platform seemed overly complicated, and the frequent changes to the portfolio as well as the messaging made us uneasy. We also decided it would not be wise to tie our application platform to a specific infrastructure provider, as Tanzu cannot be deployed on anything other than vSphere. SUSE Rancher seemed good overall, but ultimately felt closer to a DIY approach versus the comprehensive package that Red Hat OpenShift provides.
It's easy to understand what are being billed and what's included in each type of subscription. Same with the support (Std or Premium) you know exactly what to expect when you need to use it. The "core" unit approach on the subscription made really simple to scale and carry the workloads from one site to another.
This is a great platform to deployment container applications designed for multiple use cases. Its reasonably scalable platform, that can host multiple instances of applications, which can seamlessly handle the node and pod failure, if they are configured properly. There should be some scalability best practices guide would be very useful
As a DevOps Engineer , GCR has made a important contribution to my organisation because GCR can be used to store code and assets, which can help to reduce the development time for new projects.
Using GCR has been a cost-effective solution for us since we only pay for the storage we actually utilise. This has resulted in significant savings on our cloud storage expenses.
We trust GCR because they prioritise the security of our data by utilising industry-standard encryption. This assurance brings us peace of mind, as we know our information is protected.
That is a complicated question and one that's not easy for me to answer. There's a lot of factors that go into all of the stuff that we just don't have an easy way of measuring. And we realize that while we're implementing Red Hat OpenShift, we've tried to start measuring some of that stuff, but we don't have a baseline to go on. So it's hard to say. What I can tell you is general experience with the platform has been extremely positive from the development aspect. Teams have been very, very happy with the speed at which they're able to do stuff. They've been happy with that. The way it works in one environment is exactly the way it works in the next environment because we don't have configuration drift, that type of thing, and has had very positive impacts. But we didn't have a baseline to start with. So I can't talk about getting there faster or anything like that.