Amazon Redshift is a hosted data warehouse solution, from Amazon Web Services.
$0.24
per GB per month
Tableau Desktop
Score 8.4 out of 10
N/A
Tableau Desktop is a data visualization product from Tableau. It connects to a variety of data sources for combining disparate data sources without coding. It provides tools for discovering patterns and insights, data calculations, forecasts, and statistical summaries and visual storytelling.
As our applications are hosted on AWS service, Redshift is the best option for us. Also, it provide a near to real-time performance on limited datasets and less complex queries. High availability is the major concern for any growing business and AWS is the best option for this. …
Most of our stack is on AWS, so while Snowflake and BigQuery was a viable option from a performance perspective, it was easier to integrate with RedShift. We considered hosting SQL Server on AWS or using Amazon RDS (Postgres or MySQL), however, the self-service aspect of …
It's fast processing compared to other products and it's best for structured data analytics & data warehousing purpose. The unique columnar data storage architecture compared to other products makes it the great choice for analytics projects.
Amazon redshift is useful for our data sets due to its low cost and ease of use for the analysts, however, we were considering using BigQuery instead if we had chosen to pursue the Google Analytics 360 product.
Compared to Oracle Data Warehouse, Redshift is a better data warehouse. However, this comes at a cost of advanced functionality and the ability to do OLTP style processing. What you gain is faster querying time and better scalability.
Looker has the benefit of being owned by Google and seamless interface with BigQuery. We didn't use BigQuery at the time, so the benefit wasn't realized. However, we are starting to use it more and more, and will be evaluating Looker again.
In comparison to Tableau, the other dashboarding/BI tools I've used feel clunky, are very slow to develop in, and seem to lack features of a more modernized tool like Tableau. In Pentaho Analyzer, for instance, trying to include multiple worksheets or reports in a single …
If the number of connections is expected to be low, but the amounts of data are large or projected to grow it is a good solutions especially if there is previous exposure to PostgreSQL. Speaking of Postgres, Redshift is based on several versions old releases of PostgreSQL so the developers would not be able to take advantage of some of the newer SQL language features. The queries need some fine-tuning still, indexing is not provided, but playing with sorting keys becomes necessary. Lastly, there is no notion of the Primary Key in Redshift so the business must be prepared to explain why duplication occurred (must be vigilant for)
The best scenario is definitely to collect data from several sources and create dedicated dashboards for specific recipients. However, I miss the possibility of explaining these reports in more detail. Sometimes, we order a report, and after half a year, we don't remember the meaning of some data (I know it's our fault as an organization, but the tool could force better practices).
[Amazon] Redshift has Distribution Keys. If you correctly define them on your tables, it improves Query performance. For instance, we can define Mapping/Meta-data tables with Distribution-All Key, so that it gets replicated across all the nodes, for fast joins and fast query results.
[Amazon] Redshift has Sort Keys. If you correctly define them on your tables along with above Distribution Keys, it further improves your Query performance. It also has Composite Sort Keys and Interleaved Sort Keys, to support various use cases
[Amazon] Redshift is forked out of PostgreSQL DB, and then AWS added "MPP" (Massively Parallel Processing) and "Column Oriented" concepts to it, to make it a powerful data store.
[Amazon] Redshift has "Analyze" operation that could be performed on tables, which will update the stats of the table in leader node. This is sort of a ledger about which data is stored in which node and which partition with in a node. Up to date stats improves Query performance.
An excellent tool for data visualization, it presents information in an appealing visual format—an exceptional platform for storing and analyzing data in any size organization.
Through interactive parameters, it enables real-time interaction with the user and is easy to learn and get support from the community.
We've experienced some problems with hanging queries on Redshift Spectrum/external tables. We've had to roll back to and old version of Redshift while we wait for AWS to provide a patch.
Redshift's dialect is most similar to that of PostgreSQL 8. It lacks many modern features and data types.
Constraints are not enforced. We must rely on other means to verify the integrity of transformed tables.
Our use of Tableau Desktop is still fairly low, and will continue over time. The only real concern is around cost of the licenses, and I have mentioned this to Tableau and fully expect the development of more sensible models for our industry. This will remove any impediment to expansion of our use.
Just very happy with the product, it fits our needs perfectly. Amazon pioneered the cloud and we have had a positive experience using RedShift. Really cool to be able to see your data housed and to be able to query and perform administrative tasks with ease.
Tableau Desktop has proven to be a lifesaver in many situations. Once we've completed the initial setup, it's simple to use. It has all of the features we need to quickly and efficiently synthesize our data. Tableau Desktop has advanced capabilities to improve our company's data structure and enable self-service for our employees.
When used as a stand-alone tool, Tableau Desktop has unlimited uptime, which is always nice. When used in conjunction with Tableau Server, this tool has as much uptime as your server admins are willing to give it. All in all, I've never had an issue with Tableau's availability.
Tableau Desktop's performance is solid. You can really dig into a large dataset in the form of a spreadsheet, and it exhibits similarly good performance when accessing a moderately sized Oracle database. I noticed that with Tableau Desktop 9.3, the performance using a spreadsheet started to slow around 75K rows by about 60 columns. This was easily remedied by creating an extract and pushing it to Tableau Server, where performance went to lightning fast
The support was great and helped us in a timely fashion. We did use a lot of online forums as well, but the official documentation was an ongoing one, and it did take more time for us to look through it. We would have probably chosen a competitor product had it not been for the great support
Tableau support has been extremely responsive and willing to help with all of our requests. They have assisted with creating advanced analysis and many different types of custom icons, data formatting, formulas, and actions embedded into graphs. Tableau offers a weekly presentation of features and assists with internal company projects.
It is admittedly hard to train a group of people with disparate levels of ability coming in, but the software is so easy to use that this is not a huge problem; anyone who can follow simple instructions can catch up pretty quickly.
I think the training was good overall, but it was maybe stating the obvious things that a tech savvy young engineer would be able to pick up themselves too. However, the example work books were good and Tableau web community has helped me with many problems
Again, training is the key and the company provides a lot of example videos that will help users discover use cases that will greatly assist their creation of original visualizations. As with any new software tool, productivity will decline for a period. In the case of Tableau, the decline period is short and the later gains are well worth it.
Than Vertica: Redshift is cheaper and AWS integrated (which was a plus because the whole company was on AWS). Than BigQuery: Redshift has a standard SQL interface, though recently I heard good things about BigQuery and would try it out again. Than Hive: Hive is great if you are in the PB+ range, but latencies tend to be much slower than Redshift and it is not suited for ad-hoc applications.
I have used Power BI as well, the pricing is better, and also training costs or certifications are not that high. Since there is python integration in Power BI where I can use data cleaning and visualizing libraries and also some machine learning models. I can import my python scripts and create a visualization on processed data.
Redshift is relatively cheaper tool but since the pricing is dynamic, there is always a risk of exceeding the cost. Since most of our team is using it as self serve and there is no continuous tracking by a dedicated team, it really needs time & effort on analyst's side to know how much it is going to cost.
Tableau Desktop's scaleability is really limited to the scale of your back-end data systems. If you want to pull down an extract and work quickly in-memory, in my application it scaled to a few tens of millions of rows using the in-memory engine. But it's really only limited by your back-end data store if you have or are willing to invest in an optimized SQL store or purpose-built query engine like Veritca or Netezza or something similar.
Our company is moving to the AWS infrastructure, and in this context moving the warehouse environments to Redshift sounds logical regardless of the cost.
Development organizations have to operate in the Dev/Ops mode where they build and support their apps at the same time.
Hard to estimate the overall ROI of moving to Redshift from my position. However, running Redshift seems to be inexpensive compared to all the licensing and hardware costs we had on our RDBMS platform before Redshift.
Tableau was acquired years ago, and has provided good value with the content created.
Ongoing maintenance costs for the platform, both to maintain desktop and server licensing has made the continuing value questionable when compared to other offerings in the marketplace.
Users have largely been satisfied with the content, but not with the overall performance. This is due to a combination of factors including the performance of the Tableau engines as well as development deficiencies.