Google Cloud Run enables users to build and deploy scalable containerized apps written in any language (including Go, Python, Java, Node.js, .NET, and Ruby) on a fully managed platform. Cloud Run can be paired with other container ecosystem tools, including Google's Cloud Build, Cloud Code, Artifact Registry, and Docker. And it features out-of-the-box integration with Cloud Monitoring, Cloud Logging, Cloud Trace, and Error Reporting to ensure the health of an application.
N/A
IBM Cloud Kubernetes Service
Score 7.9 out of 10
Mid-Size Companies (51-1,000 employees)
IBM Cloud Kubernetes Service is a
managed Kubernetes offering, delivering user tools and built-in security for rapid delivery of applications
that users can bind to cloud services related to IBM Watson®, IoT, DevOps
and data analytics. As a certified K8s provider, IBM Cloud Kubernetes
Service provides intelligent scheduling, self-healing, horizontal
scaling, service discovery and load balancing, automated rollouts and
rollbacks, and secret and configuration management. The Kubernetes…
N/A
Red Hat OpenShift
Score 9.2 out of 10
N/A
OpenShift is Red Hat's Cloud Computing Platform as a Service (PaaS) offering. OpenShift is an application platform in the cloud where application developers and teams can build, test, deploy, and run their applications.
I have earlier used various services, such as Kubernetes and Google Cloud Run. Still, IBM Cloud Kubernetes Service offers more convenience and a larger set of functions and features are available in Kubernetes. Google Cloud Run is much closer to a server less approach to …
Verified User
Team Lead
Chose IBM Cloud Kubernetes Service
IBM Cloud is simple to use and has no custom requirements to start, Other products are not cost-efficient compared to IBM IKS.
Well, as we were in a state to compulsory implement this service, but then it is very suited for deployment of a service that has higher fault tolerance. And maintaining the lifecycle of custom made resources. And the ability to integrate other services also
Microservices and RestFul API application as it is fast and reliant. Seamless integration with event triggers such as pubsub or event arc, so you can easily integrate that with usecases with file uploads, database changes, etc. Basically great with short-lived tasks, if however, you have long-running processses, Cloud Run might not be idle for this. For example if you have a long running data processing task, other solutions such as kubeflow pipelines or dataflow are more suited for this kind of tasks. Cloud Run is also stateless, so if you need memory, you will have to connect an external database.
IBM Cloud Kubernetes Service also stands out in environments where the workloads vary continuously and require befitting scale. The product excels particularly in microservices structures, wherein the companies would harness the capacity for container orchestration and automated scaling. Still, it may face the challenges due to monolith applications that have not been originally developed for using container technology.
Red Hat OpenShift, despite its complexity and overhead, remains the most complete and enterprise-ready Kubernetes platform available. It excels in research projects like ours, where we need robust CI/CD, GPU scheduling, and tight integration with tools like Jupyter, OpenDataHub, and Quiskit. Its security, scalability, and operator ecosystem make it ideal for experimental and production-grade AI workloads. However, for simpler general hosting tasks—such as serving static websites or lightweight backend services—we find traditional VMs, Docker, or LXD more practical and resource-efficient. Red Hat OpenShift shines in complex, container-native workflows, but can be overkill for basic infrastructure needs.
IBM has a strong focus on serverless and Kubernetes. This shows in the platform. Deploying containers to Kubernetes was very easy.
Deploying a Kubernetes cluster through the GUI is very easy and quick. On top of that, IBM Cloud offers a single node cluster for Free.
Container Registry is a very good product for managing container images. Integration with Kubernetes was seemless.
Portability. To transition from Google Cloud Kubernetes to IBM Cloud Kubernetes took almost no effort. We mostly use the CLI and the standard tools such as kubectl were present.
We had a few microservices that dealt with notifications and alerts. We used OpenShift to deploy these microservices, which handle and deliver notifications using publish-subscribe models.
We had to expose an API to consumers via MTLS, which was implemented using Server secret integration in OpenShift. We were then able to deploy the APIs on OpenShift with API security.
We integrated Splunk with OpenShift to view the logs of our applications and gain real-time insights into usage, as well as provide high availability.
The UI can be made simpler. Currently the UI is bloated and it takes time to find out what you want
More integrations with container registry providers (ECR, dockerhub)
Better permissions UX. Currently GCP requires service accounts to be used with cloud products, the experience adding/removing permissions is difficult to navigate
I constantly get this error even when everything is well configured prefect.exceptions.AuthorizationError: [{'path': ['auth_info'], 'message': 'AuthenticationError: Forbidden', 'extensions': {'code': 'UNAUTHENTICATED'}}]
Then sometimes the error disapear without changine anything, happened twice to me. Should there be an issue with the authentication service? Please let's improve or let users know why this may be happening.
Improve the UX in the browse console when removing many images at once
UX on the process of installing KeyCloack operator
I wouldn't necessarily say there is look everyday technology transform. I can see a trend wherein Red Hat OpenShift is adopting all the new technology trends and helping their customers align with their priorities and the emerging technology trends. I wouldn't call out various scope for development every day. There is scope for development. It is all how the organizations adopt it and how they deliver it to their customers. I don't want to call out there is scope for development. It's happening. It is a never ending process.
At the moment, I don't have anything to call out. We are experiencing Red Hat OpenShift and we can see every day they're coming up with new features as and when they come up with new features, we want to experience it more and more. We are looking for opportunities wherein this can be leveraged to help our users and partners.
We have our application running on a CentOS compartment on IBM Cloud Kubernetes Service. We have been utilizing the help since IBM Cloud initially dispatched. We liked the adaptability and versatility that IBM Cloud Kubernetes Service give us. Since we are tiny, the Kubernetes administration is just utilized at present inside my venture bunch.
This is the current strategy for the company, most of the products in the organisation are aligning to Openshift and various use cases it support. Also lot of applications are being developed for AI use case, openshift.AI provides opportunity to host and leverage the AI capabilities for these applications
The UI/console is great... the documentation is top-notch for developers, but the CLI itself when you have to script around it is very complex and easy to forget some options... the downside of a generic command line client.
We actually haven't had any real problems in our clusters recently and the results we have gotten from adopting IBM Cloud Kubernetes Service have been beyond even our greatest expectations. The community has helped optimize the use of the system and make it relatively simpler to use.
As I said before, the obserability is one of the weakest point of OpenShift and that has a lot to do with usability. The Kibana console is not fully integrated with OpenShift console and you have to switch from tab to tab to use it. Same with Prometheus, Jaeger and Grafan, it's a "simple" integration but if you want to do complex queries or dashboards you have to go to the specific console
Redhat openshift is generally reliable and available platform, it ensures high availability for most the situations. in fact the product where we put openshift in a box, we ensure that the availability is also happening at node and network level and also at storage level, so some of the factors that are outside of Openshift realm are also working in HA manner.
Overall, this platform is beneficial. The only downsides we have encountered have been with pods that occasionally hang. This results in resources being dedicated to dead or zombie pods. Over time, these wasted resources occasionally cause us issues, and we have had difficulty monitoring these pods. However, this issue does not overshadow the benefits we get from Openshift.
The self-guided support was solid, and there are plenty of online videos to guide first time users, but I think one area of improvement is a faster way to transfer a large quantity of files from our local machine to the cloud for storage (Aspera)
Their customer support team is good and quick to respond. On a couple of occassions, they have helped us in solving some issues which we were finding a tad difficult to comprehend. On a rare occasion, the response was a bit slow but maybe it was because of the festival season. Overall a good experience on this front.
I was not involved in the in person training, so i can not answer this question, but the team in my org worked directly with Openshift and able to get the in person training done easily, i did not hear problem or complain in this space, so i hope things happen seamlessly without any issue.
Online training is really an important resource for using these tools. IBM's help center is rich in useful information and tips. Also, external guides and tutorials are available (e.g. on youtube), but I followed only IBM ones and I had no difficulties.
We went thru the training material on RH webesite, i think its very descriptive and the handson lab sesssions are very useful. It would be good to create more short duration videos covering one single aspect of openshift, this wll keep the interest and also it breaks down the complexity to reasonable chunks.
Ease of use. Very intuitive. We have been looking for a product that allows us to orchestrate our docker containers in a way where it allows us to effectively scale our applications to production. It also provides us a way of monitoring all our infrastructure in a very clear concise way.
We mainly selected [IBM Cloud Kubernetes Service] because IBM fabric blockchain service is mostly compatible with it. To have all the infrastructure in a single cloud to get the best output we selected the [IBM Cloud Kubernetes Service].
The Tanzu Platform seemed overly complicated, and the frequent changes to the portfolio as well as the messaging made us uneasy. We also decided it would not be wise to tie our application platform to a specific infrastructure provider, as Tanzu cannot be deployed on anything other than vSphere. SUSE Rancher seemed good overall, but ultimately felt closer to a DIY approach versus the comprehensive package that Red Hat OpenShift provides.
It's easy to understand what are being billed and what's included in each type of subscription. Same with the support (Std or Premium) you know exactly what to expect when you need to use it. The "core" unit approach on the subscription made really simple to scale and carry the workloads from one site to another.
IBM's CKS does not offers automatic autoscaling nor vertical scaling (automatic). Other services like Google Kubernetes Engine scales up and down very well
This is a great platform to deployment container applications designed for multiple use cases. Its reasonably scalable platform, that can host multiple instances of applications, which can seamlessly handle the node and pod failure, if they are configured properly. There should be some scalability best practices guide would be very useful
That is a complicated question and one that's not easy for me to answer. There's a lot of factors that go into all of the stuff that we just don't have an easy way of measuring. And we realize that while we're implementing Red Hat OpenShift, we've tried to start measuring some of that stuff, but we don't have a baseline to go on. So it's hard to say. What I can tell you is general experience with the platform has been extremely positive from the development aspect. Teams have been very, very happy with the speed at which they're able to do stuff. They've been happy with that. The way it works in one environment is exactly the way it works in the next environment because we don't have configuration drift, that type of thing, and has had very positive impacts. But we didn't have a baseline to start with. So I can't talk about getting there faster or anything like that.