IBM Cognos is a full-featured business intelligence suite by IBM, designed for larger deployments. It comprises Query Studio, Reporting Studio, Analysis Studio and Event Studio, and Cognos Administration along with tools for Microsoft Office integration, full-text search, and dashboards.
$10
per month per user
JMP
Score 9.1 out of 10
N/A
JMP® is statistical analysis software with capabilities that span from data access to advanced statistical techniques, with click of a button sharing. The software is interactive and visual, and statistically deep enough to allow users to see and explore data.
IBM Cognos Analytics has great scheduling capabilities. A single report can be parameterized (e.g., “Store Manager ID”) and burst to thousands of recipients with their slice of data.IBM Cognos Analytics is a good fit for highly complex, multi-level calculations which can be handled by Report Studio. For example Monthly balance sheet that requires multi step calculation
It is perfectly suited for statistical analyses, but I would not recommend JMP for users who do not have a statistical background. As previously stated, the learning curve is exceptionally steep, and I think that it would prove to be too steep for those without statistical background/knowledge
JMP is designed from the ground-up to be a tool for analysts who do not have PhDs in Statistics without in anyway "dumbing down" the level of statistical analysis applied. In fact, JMP operationalizes the most advanced statistical methods. JMP's design is centred on the JMP data table and dialog boxes. It is data focused not jargon-focussed. So, unlike other software where you must choose the correct statistical method (eg. contingency, ANOVA, linear regression, etc.), with JMP you simply assign the columns in a dialog into roles in the analysis and it chooses the correct statistical method. It's a small thing but it reflects the thinking of the developers: analysts know their data and should only have to think about their data. Analyses should flow from there.
JMP makes most things interactive and visual. This makes analyses dynamic and engaging and obviates the complete dependence on understanding p-values and other statistical concepts(though they are all there) that are often found to be foreign or intimidating.
One of the best examples of this is JMP's profiler. Rather than looking at static figures in a spreadsheet, or a series of formulas, JMP profiles the formulas interactively. You can monitor the effect of changing factors (Xs) and see how they interact with other factors and the responses. You can also specify desirability (maximize, maximize, match-target) and their relative importances to find factor settings that are optimal. I have spent many lengthy meetings working with the profiler to review design and process options with never a dull moment.
The design of experiments (DOE) platform is simply outstanding and, in fact, the principal developers of it have won several awards. Over the last 15 years, using methods broadly known as an "exchange algorithm," JMP can create designs that are far more flexible than conventional designs. This means, for example, that you can create a design with just the interactions that are of interest; you can selectively choose those interactions that are not of interest and drop collecting their associated combinations.
Classical designs are rigid. For example, a Box-Benhken or other response surface design can have only continuous factors. What if you want to investigate these continuous factors along with other categorical factors such as different categorical variables such as materials or different furnace designs and look at the interaction among all factors? This common scenario cannot be handled with conventional designs but are easily accommodated with JMP's Custom DOE platform.
The whole point of DOE is to be able to look at multiple effects comprehensively but determine each one's influence in near or complete isolation. The custom design platform, because it produces uniques designs, provides the means to evaluate just how isolated the effects are. This can be done before collecting data because this important property of the DOE is a function of the design, not the data. By evaluating these graphical reports of the quality of the design, the analyst can make adjustments, adding or reducing runs, to optimize cost, effort and expected learnings.
Over the last number of releases of JMP, which appear about every 18 months now, they have skipped the dialog boxes to direct, drag-and-drop analyses for building graphs and tables as well as Statistical Process Control Charts. Interactivity such as this allows analysts to "be in the moment." As with all aspects of JMP, they are thinking of their subject matter without the cumbersomeness associated with having to think about statistical methods. It's rather like a CEO thinking about growing the business without having to think about every nuance and intricacy of accounting. The statistical thinking is burned into the design of JMP.
Without data analysis is not possible. Getting data into a situation where it can be analyzed can be a major hassle. JMP can pull data from a variety of sources including Excel spreadsheets, CSV, direct data feeds and databases via ODBC. Once the data is in JMP it has all the expected data manipulation capabilities to form it for analysis.
Back in 2000 JMP added a scripting language (JMP Scripting Language or JSL for short) to JMP. With JSL you can automate routine analyses without any coding, you can add specific analyses that JMP does not do out of the box and you can create entire analytical systems and workflows. We have done all three. For example, one consumer products company we are working with now has a need for a variant of a popular non-parametric analysis that they have employed for years. This method will be found in one of the menus and appear as if it were part of JMP to begin with. As for large systems, we have written some that are tens of thousands of lines that take the form of virtual labs and process control systems among others.
JSL applications can be bundled and distributed as JMP Add-ins which make it really easy for users to add to their JMP installation. All they need to do is double-click on the add-in file and it's installed. Pharmaceutical companies and others who are regulated or simply want to control the JMP environment can lock-down JMP's installation and prevent users from adding or changing functionality. Here, add-ins can be distributed from a central location that is authorized and protected to users world-wide.
JMP's technical support is second to none. They take questions by phone and email. I usually send email knowing that I'll get an informed response within 24 hours and if they cannot resolve a problem they proactively keep you informed about what is being done to resolve the issue or answer your question.
IBM Cognos Analytics enables customer data segmentation, which is essential for marketing, improving and streamlining purchasing behavior and preferences. This helps companies create more targeted and effective marketing campaigns.
Our clients Through data analysis, we can identify and observe trends in the behavior of other clients, allowing us to anticipate needs and adjust strategies to avoid consequences.
In general JMP is much better fit for a general "data mining" type application. If you want a specific statistics based toolbox, (meaning you just want to run some predetermined test, like testing for a different proportion) then JMP works, but is not the best. JMP is much more suited to taking a data set and starting from "square 1" and exploring it through a range of analytics.
The CPK (process capability) module output is shockingly poor in JMP. This sticks out because, while as a rule everything in JMP is very visual and presentable, the CPK graph is a single-line-on-grey-background drawing. It is not intuitive, and really doesn't tell the story. (This is in contrast with a capability graph in Minitab, which is intuitive and tells a story right off.) This is also the case with the "guage study" output, used for mulivary analysis in a Six Sigma project. It is not intuitive and you need to do a lot of tweaking to make the graph tell you the story right off. I have given this feedback to JMP, and it is possible that it will be addressed in future versions.
I've never heard of JMP allowing floating licenses in a company. This will ALWAYS be a huge sticking point for small to middle size companies, that don't have teams people dedicated to analytics all day. If every person that would do problem solving needs his/her own seat, the cost can be prohibitive. (It gets cheaper by the seat as you add licenses, but for a small company that might get no more than 5 users, it is still a hard sell.)
For an existing solution, renewing licenses does provide a good return on investment. Additionally, while rolling out scorecards and dashboards with little adhoc capabilities, to end users, cognos is very easily scalable. It also allows to create a solution that has a mix of OLAP and relational data-sources, which is a limitation with other tools. Synchronizing with existing security setup is easy too.
JMP has been good at releasing updates and adding new features and their support is good. Analytics is quick and you don't need scripting/programming experience. It has been used organization wide, and works well in that respect. Open source means that there are concerns regarding timely support. Cheap licensing and easy to maintain.
We have a strong user base (3500 users) that are highly utilizing this tool. Basic users are able to consume content within the applied security model. We have a set of advanced users that really push the limits of Cognos with Report and Query Studio. These users have created a lot of personal content and stored it in 'My Reports'. Users enjoy this flexibility.
The GUI interface makes it easier to generate plots and find statistics without having to write code. The JSL scripting is a bit of a steep learning curve but does give you more ability to customize your analysis. Overall, I would recommend JMP as a good product for overall usability.
Reports can typically be viewed through any browser that can access the server, so the availability is ultimately up to what the company utilizing it is comfortable with allowing, though report development tends to be more picky about browsers and settings as mentioned above. It also has an optional iPad app and general mobile browsing support, but dashboards lack the mobile compatibility. What keeps it from getting a higher score is the desktop tools that are vital to the development process. The compatibility with only Windows when the server has a wide range of compatibility can be a real sore point for a company that outfits its employees exclusively with Mac or Linux machines. Of course, if they are planning on outsourcing the development anyways, it's a rather moot point
Overall no major complaints but it doesn't handle DMR (Dimensionally Modeled for Relational) very well. DMR modelling is a capability that IBM Cognos Framework Manager provides allowing you to specify dimensional information for relational metadata and allows for OLAP-style queries. However, the capability is not very efficient and, for example, if I'm using only 2 columns on a 20-column model, the software is not smart enough to exclude 18 columns and the query side gets progressively larger and larger until it's effectively unusable.
Why is their web application not working as fast as you think it should? They never know, and it is always a a bunch of shots in the dark to find out. Trying to download software from them is like trying to find a book at the library before computers were invented.
Support is great and give ease of contact, rapid response, and willingness to 'stick to the task' until resolution or acknowledgement that the problem would have to be resolved in a future build. Basically, one gets the very real sense that another human being is sensitive to your problems - great or small.
Onsite training provided by IBM Cognos was effective and as expected. They did not perform training with our data which was a bit difficult for our end-users.
The online courses they offer are thorough and presented in such a way that someone who isn't already familiar with the general design methodologies used in this field will be capable of making a good design. The training environments are provided as a fully self contained virtual machine with everything needed already to create the environments. We've had some persisting issues with the environments becoming unavailable, but support has been responsive when these issues arise and straightening them out for us
Make sure that any custom tables that you have, are built into your metadata packages. You can still access them via SQL queries in Cognos, but it is much easier to have them as a part of the available metadata packages.
My company selected IBM Congos Analytics because of its advanced features and data representation for data analysis. Its row and column features are very effective for creating dashboards and reports to visualize data. It's chart representation and view format are very attractive and useful for representation.
MS Excel with AnalysisToolPak provides a home-grown solution, but requires a high degree of upkeep and is difficult to hand off. Minitab is the closes competitor, but JMP is better suited to the production environment, roughly equivalent in price, and has superior support.
The Cognos architecture is well suited for scalability. However, the architecture must be designed with scalability in mind from day one of the implementation. We recently upgraded from 10.1 to 10.2.1 and took the opportunity to revamp our architecture. It is now poised for future growth and scalability.
ROI: Even if the cost can be high, the insights you get out of the tool would definitely be much more valuable than the actual cost of the software. In my case, most of the results of your analysis were shown to the client, who was blown away, making the money spent well worth for us.
Potential negative: If you are not sure your team will use it, there's a chance you will just waste money. Sometimes the IT department (usually) tries to deploy a better tool for the entire organization but they keep using the old tool they are used too (most likely MS Excel).