Astra DB is a vector database for developers. In 2025 Datastax, the developer and supporter of Astra DB, was acquired. Astra DB is now available as a component of the IBM watsonx.data Multicloud offering.
N/A
Google BigQuery
Score 8.7 out of 10
N/A
Google's BigQuery is part of the Google Cloud Platform, a database-as-a-service (DBaaS) supporting the querying and rapid analysis of enterprise data.
$6.25
per TiB (after the 1st 1 TiB per month, which is free)
MySQL
Score 8.3 out of 10
N/A
MySQL is a popular open-source relational and embedded database, now owned by Oracle.
I have previously used and evaluated MongoDB and MySQL for various projects before choosing AstraDB for my chatbot application. While MongoDB and MySQL are both powerful and popular database solutions, AstraDB stood out for specific reasons in the context of my project.MongoDB, …
The biggest competitor was Cassandra which we have been using as the self-hosted solution, so we had the option of going to hosted versions as well. The main advantage of Astra was the ability to combine managed experience with scalability, which was Datastax' strong suit. The …
Compared to PostgreSQL and MySQL, Google BigQuery is faster and more scalable for large datasets. It’s serverless, so there’s no need to manage infrastructure. We chose Google BigQuery for its ease of use built-in analytics features
Main reason is how it integrates directly with the google ecosystem which really facilitates the automatization proceses for the whole company. This ensures that sales and all the other departments have the correct information on a daily bases with a ease of use with day to day …
I have used most of the data analytics platforms. Based on my work, I have found that the user interface of Google BigQuery is simple to navigate. I like the front view - ease of joining tables, and integration with other platforms.
I have used Google BigQuery and it is very difficult to start with it. Although it is very fast and the speed performance is much better with BigQuery but it costs and is very difficult to start with. There's also no proper documentation on it, so MySQL wins in terms of …
Before MySQL, our team was exploring and evaluating different options for a good RDMS (relational database management system) service. We explored Oracle, MSSQL, and Google BigQuery. Most of these are costly and not easy to maintain in the long run in terms of price especially …
MySQL has most of the functionality of other, very costly, alternatives without the big price tag. It is open-source with improvements coming at a relatively good rate. It is not as robust as those other offerings and can have some challenging points at scale for large …
We've been super happy with Astra DB. It's been extremely well-suited for our vector search needs as described in previous responses. With Astra DB’s high-performance vector search, Maester’s AI dynamically optimizes responses in real-time, adapting to new user interactions without requiring costly retraining cycles.
Event-based data can be captured seamlessly from our data layers (and exported to Google BigQuery). When events like page-views, clicks, add-to-cart are tracked, Google BigQuery can help efficiently with running queries to observe patterns in user behaviour. That intermediate step of trying to "untangle" event data is resolved by Google BigQuery. A scenario where it could possibly be less appropriate is when analysing "granular" details (like small changes to a database happening very frequently).
MySQL is best suited for applications on platform like high-traffic content-driven websites, small-scale web apps, data warehouses which regards light analytical workloads. However its less suited for areas like enterprise data warehouse, OLAP cubes, large-scale reporting, applications requiring flexible or semi-structured data like event logging systems, product configurations, dynamic forms.
We need to be able to process a lot of data (our biggest clients process hundreds of milions of transactions every month). However, it is not only the amount of data, it is also an unpredictable patterns with spikes occuring at different points of time - something athat Astra is great at.
Our processing needs to be extremaly fast. Some of our clients use our enrichment in a synchronous way, meaning that any delay in processing is holding up the whole transaction lifecycle and can have a major impact on the client. Astra is very fast.
A close collaboration with GCP makes our life very easy. All of our technology sits in Google Cloud, so having Astra in there makes it a no-brainer solution for us.
GSheet data can be linked to a BigQuery table and the data in that sheet is ingested in realtime into BigQuery. It's a live 'sync' which means it supports insertions, deletions, and alterations. The only limitation here is the schema'; this remains static once the table is created.
Seamless integration with other GCP products.
A simple pipeline might look like this:-
GForms -> GSheets -> BigQuery -> Looker
It all links up really well and with ease.
One instance holds many projects.
Separating data into datamarts or datameshes is really easy in BigQuery, since one BigQuery instance can hold multiple projects; which are isolated collections of datasets.
The support team sometimes requires the escalate button pressed on tickets, to get timely responses. I will say, once the ticket is escalated, action is taken.
They require better documentation on the migration of data. The three primary methods for migrating large data volumes are bulk, Cassandra Data Migrator, and ZDM (Zero Downtime Migration Utility). Over time I have become very familiar will all three of these methods; however, through working with the Services team and the support team, it seemed like we were breaking new ground. I feel if the utilities were better documented and included some examples and/or use cases from large data migrations; this process would have been easier. One lesson learned is you likely need to migrate your application servers to the same cloud provider you host Astra on; otherwise, the latency is too large for latency-sensitive applications.
Please expand the availability of documentation, tutorials, and community forums to provide developers with comprehensive support and guidance on using Google BigQuery effectively for their projects.
If possible, simplify the pricing model and provide clearer cost breakdowns to help users understand and plan for expenses when using Google BigQuery. Also, some cost reduction is welcome.
It still misses the process of importing data into Google BigQuery. Probably, by improving compatibility with different data formats and sources and reducing the complexity of data ingestion workflows, it can be made to work.
Learning curve: is big. Newbies will face problems in understanding the platform initially. However, with plenty of online resources, one can easily find solutions to problems and learn on the go.
Backup and restore: MySQL is not very seamless. Although the data is never ruptured or missed, the process involved is not very much user-friendly. Maybe, a new command-line interface for only the backup-restore functionality shall be set up again to make this very important step much easier to perform and maintain.
We have to use this product as its a 3rd party supplier choice to utilise this product for their data side backend so will not be likely we will move away from this product in the future unless the 3rd party supplier decides to change data vendors.
For teaching Databases and SQL, I would definitely continue to use MySQL. It provides a good, solid foundation to learn about databases. Also to learn about the SQL language and how it works with the creation, insertion, deletion, updating, and manipulation of data, tables, and databases. This SQL language is a foundation and can be used to learn many other database related concepts.
I think overall it is easy to use. I haven't done anything from the development side but an more of an end user of reporting tables built in Google BigQuery. I connect data visualization tools like Tableau or Power BI to the BigQuery reporting tables to analyze trends and create complex dashboards.
I give MySQL a 9/10 overall because I really like it but I feel like there are a lot of tech people who would hate it if I gave it a 10/10. I've never had any problems with it or reached any of its limitations but I know a few people who have so I can't give it a 10/10 based on those complaints.
I have never had any significant issues with Google Big Query. It always seems to be up and running properly when I need it. I cannot recall any times where I received any kind of application errors or unplanned outages. If there were any they were resolved quickly by my IT team so I didn't notice them.
I think Google Big Query's performance is in the acceptable range. Sometimes larger datasets are somewhat sluggish to load but for most of our applications it performs at a reasonable speed. We do have some reports that include a lot of complex calculations and others that run on granular store level data that so sometimes take a bit longer to load which can be frustrating.
Their response time is fast, in case you do not contact them during business hours, they give a very good follow-up to your case. They also facilitate video calls if necessary for debugging.
BigQuery can be difficult to support because it is so solid as a product. Many of the issues you will see are related to your own data sets, however you may see issues importing data and managing jobs. If this occurs, it can be a challenge to get to speak to the correct person who can help you.
We have never contacted MySQL enterprise support team for any issues related to MySQL. This is because we have been using primarily the MySQL Server community edition and have been using the MySQL support forums for any questions and practical guidance that we needed before and during the technical implementations. Overall, the support community has been very helpful and allowed us to make the most out of the community edition.
Graph, search, analytics, administration, developer tooling, and monitoring are all incorporated into a single platform by Astra DB. Mongo Db is a self-managed infrastructure. Astra DB has Wide column store and Mongo DB has Document store. The best thing is that Astra DB operates on Java while Mongo DB operates on C++
PowerBI can connect to GA4 for example but the data processing is more complicated and it takes longer to create dashboards. Azure is great once the data import has been configured but it's not an easy task for small businesses as it is with BigQuery.
MongoDB has a dynamic schema for how data is stored in 'documents' whereas MySQL is more structured with tables, columns, and rows. MongoDB was built for high availability whereas MySQL can be a challenge when it comes to replication of the data and making everything redundant in the event of a DR or outage.
We are well aware of the Cassandra architecture and familiar with the open source tooling that Datastax provides the industry (K8sSandra / Stargate) to scale Cassandra on Kubernetes.
Having prior knowledge of Cassandra / Kubernetes means we know that under the hood Astra is built on infinitely scalable technologies. We trust that the foundations that Astra is built on will scale so we know Astra will scale.
We have continued to expand out use of Google Big Query over the years. I'd say its flexibility and scalability is actually quite good. It also integrates well with other tools like Tableau and Power BI. It has served the needs of multiple data sources across multiple departments within my company.
Google Support has kindly provide individual support and consultants to assist with the integration work. In the circumstance where the consultants are not present to support with the work, Google Support Helpline will always be available to answer to the queries without having to wait for more than 3 days.
Previously, running complex queries on our on-premise data warehouse could take hours. Google BigQuery processes the same queries in minutes. We estimate it saves our team at least 25% of their time.
We can target our marketing campaigns very easily and understand our customer behaviour. It lets us personalize marketing campaigns and product recommendations and experience at least a 20% improvement in overall campaign performance.
Now, we only pay for the resources we use. Saved $1 million annually on data infrastructure and data storage costs compared to our previous solution.