CREE8 is a cloud SaaS platform for remote and virtual workflows for all production activities. It gives global teams access to cloud CPU/GPU and storage to deliver today’s demanding creative projects, in real time.
N/A
IBM Cloud Managed Istio
Score 8.7 out of 10
N/A
The IBM Cloud Kubernetes Service provides the Managed Istio installation add on, designed to provide additonal control over clusters and the microservices they comprise via automatic updates and lifecycle management of control plane components, and integration with platform logging and monitoring tools.
N/A
Red Hat OpenShift
Score 9.2 out of 10
N/A
OpenShift is Red Hat's Cloud Computing Platform as a Service (PaaS) offering. OpenShift is an application platform in the cloud where application developers and teams can build, test, deploy, and run their applications.
If you have contractors or workers working off site on multiple work stations, it's easy to not carry a laptop everywhere. Just log in and use the Bebop software on the cloud.
Clearly, the [IBM Cloud Managed Istio] tool is very useful when you have multiple services and each service is connecting with other services through APIs in different networks. To manage this type of complex network, [IBM Cloud Managed Istio] is very useful. It comes with a license that can increase the billing of your project so make sure if your application network mesh, monitoring cannot be managed on your own then you can use it. If your application is not very complex then you have many tools available like Grafana, Prometheus, Sumo Logic, which you can integrate individually with your cluster and implement. In this type of scenario, it is better to not use [IBM Cloud Managed Istio] and it will serve your purpose as well.
Red Hat OpenShift, despite its complexity and overhead, remains the most complete and enterprise-ready Kubernetes platform available. It excels in research projects like ours, where we need robust CI/CD, GPU scheduling, and tight integration with tools like Jupyter, OpenDataHub, and Quiskit. Its security, scalability, and operator ecosystem make it ideal for experimental and production-grade AI workloads. However, for simpler general hosting tasks—such as serving static websites or lightweight backend services—we find traditional VMs, Docker, or LXD more practical and resource-efficient. Red Hat OpenShift shines in complex, container-native workflows, but can be overkill for basic infrastructure needs.
We had a few microservices that dealt with notifications and alerts. We used OpenShift to deploy these microservices, which handle and deliver notifications using publish-subscribe models.
We had to expose an API to consumers via MTLS, which was implemented using Server secret integration in OpenShift. We were then able to deploy the APIs on OpenShift with API security.
We integrated Splunk with OpenShift to view the logs of our applications and gain real-time insights into usage, as well as provide high availability.
I wouldn't necessarily say there is look everyday technology transform. I can see a trend wherein Red Hat OpenShift is adopting all the new technology trends and helping their customers align with their priorities and the emerging technology trends. I wouldn't call out various scope for development every day. There is scope for development. It is all how the organizations adopt it and how they deliver it to their customers. I don't want to call out there is scope for development. It's happening. It is a never ending process.
At the moment, I don't have anything to call out. We are experiencing Red Hat OpenShift and we can see every day they're coming up with new features as and when they come up with new features, we want to experience it more and more. We are looking for opportunities wherein this can be leveraged to help our users and partners.
This is the current strategy for the company, most of the products in the organisation are aligning to Openshift and various use cases it support. Also lot of applications are being developed for AI use case, openshift.AI provides opportunity to host and leverage the AI capabilities for these applications
Bebop is very easy to use. As long as you have a decent computer hardware and a fast internet connection, you can access programs anywhere, any time. This makes it very convenient. In addition, I don't have to carry my computer everywhere. I leave one at the office and one at home.
As I said before, the obserability is one of the weakest point of OpenShift and that has a lot to do with usability. The Kibana console is not fully integrated with OpenShift console and you have to switch from tab to tab to use it. Same with Prometheus, Jaeger and Grafan, it's a "simple" integration but if you want to do complex queries or dashboards you have to go to the specific console
Redhat openshift is generally reliable and available platform, it ensures high availability for most the situations. in fact the product where we put openshift in a box, we ensure that the availability is also happening at node and network level and also at storage level, so some of the factors that are outside of Openshift realm are also working in HA manner.
Overall, this platform is beneficial. The only downsides we have encountered have been with pods that occasionally hang. This results in resources being dedicated to dead or zombie pods. Over time, these wasted resources occasionally cause us issues, and we have had difficulty monitoring these pods. However, this issue does not overshadow the benefits we get from Openshift.
Their customer support team is good and quick to respond. On a couple of occassions, they have helped us in solving some issues which we were finding a tad difficult to comprehend. On a rare occasion, the response was a bit slow but maybe it was because of the festival season. Overall a good experience on this front.
I was not involved in the in person training, so i can not answer this question, but the team in my org worked directly with Openshift and able to get the in person training done easily, i did not hear problem or complain in this space, so i hope things happen seamlessly without any issue.
We went thru the training material on RH webesite, i think its very descriptive and the handson lab sesssions are very useful. It would be good to create more short duration videos covering one single aspect of openshift, this wll keep the interest and also it breaks down the complexity to reasonable chunks.
The Tanzu Platform seemed overly complicated, and the frequent changes to the portfolio as well as the messaging made us uneasy. We also decided it would not be wise to tie our application platform to a specific infrastructure provider, as Tanzu cannot be deployed on anything other than vSphere. SUSE Rancher seemed good overall, but ultimately felt closer to a DIY approach versus the comprehensive package that Red Hat OpenShift provides.
It's easy to understand what are being billed and what's included in each type of subscription. Same with the support (Std or Premium) you know exactly what to expect when you need to use it. The "core" unit approach on the subscription made really simple to scale and carry the workloads from one site to another.
This is a great platform to deployment container applications designed for multiple use cases. Its reasonably scalable platform, that can host multiple instances of applications, which can seamlessly handle the node and pod failure, if they are configured properly. There should be some scalability best practices guide would be very useful
That is a complicated question and one that's not easy for me to answer. There's a lot of factors that go into all of the stuff that we just don't have an easy way of measuring. And we realize that while we're implementing Red Hat OpenShift, we've tried to start measuring some of that stuff, but we don't have a baseline to go on. So it's hard to say. What I can tell you is general experience with the platform has been extremely positive from the development aspect. Teams have been very, very happy with the speed at which they're able to do stuff. They've been happy with that. The way it works in one environment is exactly the way it works in the next environment because we don't have configuration drift, that type of thing, and has had very positive impacts. But we didn't have a baseline to start with. So I can't talk about getting there faster or anything like that.