1) Securing your back-end APIs - If you have a legacy back-end web service that has a basic authentication scheme, you can add some additional security by placing APIM in front, and requiring subscription keys. Leverage your existing firewall to ensure only your APIM instance can communicate with your back-end API, and you've basically added a layer of protection.
2) Lift and shift - there are always going to be clients that don't want to update their clients to use a newer API; in some cases you can make a newer API look like an older one by implementing some complex policies in APIM. You can also do the opposite, making older APIs look new, such as making an XML back-end accept both JSON and XML.
3) Centralizing your APIs - if you've acquired another company and want to make their API set look as if it's a part of the larger whole, APIM is an easy way to provide a consistent front-end interface for developers.
For an organisation that has completely adopted SAFe structure including naming terminology, it is less appropriate and apart from that. It can suit any organisation out there, and it can solve all your problems one way or another by customising it. It is a robust and highly scalable solution to support all the business needs. It improves a lot of productivity and visibility.
If you have a mix of automation & manual test suites, HPALM is the best tool to manage that. It definitely integrates very well with HP automation tools like HP Unified Functional Testing and HP LoadRunner. Automated Suites can be executed, reports can be maintained automatically. It also classifies which test suites are manual & which are automated & managers can see the progress happening in moving from manual to automated suites. In HPA ALM all the functional test suites, performance test suites, security suites can be defined, managed & tracked in one place.
It is a wonderful tool for test management. Whether you want to create test cases, or import it, from execution to snapshot capturing, it supports all activities very well. The linking of defects to test runs is excellent. Any changes in mandatory fields or status of the defect triggers an e-mail and sent automatically to the user that the defect is assigned to.
It also supports devops implementation by interacting with development tool sets such as Jenkins & GIT. It also bring in team collaboration by supporting collaboration tools like Slack and Hubot.
This tool can integrate to any environment, any source control management tool bringing in changes and creates that trace-ability and links between source control changes to requirements to tests across the sdlc life-cycle.
Lack of robustness is a bit of an issue. Several other providers offer more options and capabilities, but then, they are lacking in interface ease.
As with anything Azure, pricing is really hard to stay on top of. I always find that you really don’t know what you’re paying for until you get the bill. Having an excellent Azure Administrator can help resolve that.
Integrating with app services outside of Azure can be a challenge, or at least much more challenging than just using Azure App Services.
The requirements module is not as user friendly as other applications, such as Blue Bird. Managing requirements is usually done in another tool. However, having the requirements in ALM is important to ensure traceability to tests and defects.
Reporting across multiple ALM repositories is not supported within the tool. Only graphs are included within ALM functionality. Due to size considerations, one or two projects is not a good solution. Alternatively, we have started leveraging the template functionality within ALM and are integrating with a third party reporting tool to work around this issue.
NET (not Octane) requires a package for deployment to machines without administrative rights. Every time there is a change, a new package must be created, which increases the time to deploy. It also forces us to wait until multiple patches have been provided before updating production.
Because it lets me track the test cases with detailed scenarios and is clearly separated in folders. Also the defect filter helps me filter only the ones that have been assigned to a particular area of interest. The availability of reports lets me see the essentials fields which I might be missing the data on and helps me to work on these instead of having to go through everything.
It is a great tool, however, it got this rating because there is a lot of learning that takes a lot longer than other tools. There are no mobile versions of ALM even with just a project summary view. I believe ALM is well capable of integration with other analytics tools that can help business solutions prediction based on current and past project data. This is Data held in ALM but with no other use apart from human reading and project progress. ALM looks like a steady platform that I believe can handle more dynamic functionality. You could add an internal communication platform that is not a third party. Limit that communication tool to specific project members.
We have other tools in our organization like Atlassian JIRA and Microsoft Team Foundation Server, which are very capable tools but very narrow in their approach and feature set and does not come even close to the some of the core capabilities of HP ALM. HP ALM is the "System of Record" in our organization. It gives visibility for an artifact throughout the delivery chain, which cut downs unnecessary bottlenecks and noise during releases.