PowerBI can connect to GA4 for example but the data processing is more complicated and it takes longer to create dashboards. Azure is great once the data import has been configured but it's not an easy task for small businesses as it is with BigQuery.
Google BigQuery i would say is better to use than AWS Redshift but not SQL products but this could be due to being more experience in Microsoft and AWS products. It would be really nice if it could use standard SQL server coding rather than having to learn another dialect of …
Google BigQuery integrates seamlessly with Web Analytics data compared to the Azure cloud. Google BigQuery integrates natively with different digital media platforms compared to Azure and AWs.
Well suited for: Games, Chat rooms, real time software like corporate events, marathons and so. Anytime and anywhere you could use a NoSQL DB you should think of SimpleDB.
As an arduous AWS user, Amazon SimpleDB easily integrates with EC2 and other AWS module; and if you are not an AWS user, you also have a fantastic tool that will solve the problem for which you are focused.
Event-based data can be captured seamlessly from our data layers (and exported to Google BigQuery). When events like page-views, clicks, add-to-cart are tracked, Google BigQuery can help efficiently with running queries to observe patterns in user behaviour. That intermediate step of trying to "untangle" event data is resolved by Google BigQuery. A scenario where it could possibly be less appropriate is when analysing "granular" details (like small changes to a database happening very frequently).
We have found it's a great alternative for making older legacy applications work with online databases instead of only on-premises databases. We've converted over a dozen applications this way, and it has allowed our clients to have a distributed workforce using their applications without incurring the expense of a complete application rewrite.
GSheet data can be linked to a BigQuery table and the data in that sheet is ingested in realtime into BigQuery. It's a live 'sync' which means it supports insertions, deletions, and alterations. The only limitation here is the schema'; this remains static once the table is created.
Seamless integration with other GCP products.
A simple pipeline might look like this:-
GForms -> GSheets -> BigQuery -> Looker
It all links up really well and with ease.
One instance holds many projects.
Separating data into datamarts or datameshes is really easy in BigQuery, since one BigQuery instance can hold multiple projects; which are isolated collections of datasets.
Maintenance is always an issue, so using a cloud solution saves a lot of trouble.
On premise solutions always suffer from fragmented implementations here and there, where several "dba's" keep track of security and maintenance. With a cloud database it's much easier to keep a central overview.
Security options in SQL database are next level... data masking, hiding sensitive data where always neglected on premise, whereas you'll get this automatically in the cloud.
Please expand the availability of documentation, tutorials, and community forums to provide developers with comprehensive support and guidance on using Google BigQuery effectively for their projects.
If possible, simplify the pricing model and provide clearer cost breakdowns to help users understand and plan for expenses when using Google BigQuery. Also, some cost reduction is welcome.
It still misses the process of importing data into Google BigQuery. Probably, by improving compatibility with different data formats and sources and reducing the complexity of data ingestion workflows, it can be made to work.
One needs to be aware that some T-SQL features are simply not available.
The programmatic access to server, trace flags, hardware from within Azure SQL Database is taken away (for a good reason).
No SQL Agent so your jobs need to be orchestrated differently.
The maximum concurrent logins maybe an unexpected problem.
Sudden disconnects.
The developers and admin must study the capacity and tier usage limits https://docs.microsoft.com/en-us/azure/azure-subscription-service-limits otherwise some errors or even transaction aborts never seen before can occur.
Only one Latin Collation choice.
There is no way to debug T-SQL ( a big drawback in my point of view).
We have to use this product as its a 3rd party supplier choice to utilise this product for their data side backend so will not be likely we will move away from this product in the future unless the 3rd party supplier decides to change data vendors.
I think overall it is easy to use. I haven't done anything from the development side but an more of an end user of reporting tables built in Google BigQuery. I connect data visualization tools like Tableau or Power BI to the BigQuery reporting tables to analyze trends and create complex dashboards.
The interfaces are intuitive once you are familiar with all the functions. The ability to use different tools to interact with the platform, such as directly via a browser or code editors such as VS Code or Visual Studio is a great option and allows for integrating withn the project and other testing and developing tools.
I have never had any significant issues with Google Big Query. It always seems to be up and running properly when I need it. I cannot recall any times where I received any kind of application errors or unplanned outages. If there were any they were resolved quickly by my IT team so I didn't notice them.
I think Google Big Query's performance is in the acceptable range. Sometimes larger datasets are somewhat sluggish to load but for most of our applications it performs at a reasonable speed. We do have some reports that include a lot of complex calculations and others that run on granular store level data that so sometimes take a bit longer to load which can be frustrating.
BigQuery can be difficult to support because it is so solid as a product. Many of the issues you will see are related to your own data sets, however you may see issues importing data and managing jobs. If this occurs, it can be a challenge to get to speak to the correct person who can help you.
We give the support a high rating simply because every time we've had issues or questions, representatives were in contact with us quickly. Without fail, our issues/questions were handled in a timely matter. That kind of response is integral when client data integrity and availability is in question. There is also a wealth of documentation for resolving issues on your own.
It integrates beautifully with AWS. In some projects we use SimpleDB while we use DynamoDB for others, according to the characteristics of the project. If the infrastructure is AWS, we always think of one of them.
PowerBI can connect to GA4 for example but the data processing is more complicated and it takes longer to create dashboards. Azure is great once the data import has been configured but it's not an easy task for small businesses as it is with BigQuery.
We moved away from Oracle and NoSQL because we had been so reliant on them for the last 25 years, the pricing was too much and we were looking for a way to cut the cord. Snowflake is just too up in the air, feels like it is soon to be just another line item to add to your Azure subscription. Azure was just priced right, easy to migrate to and plenty of resources to hire to support/maintain it. Very easy to learn, too.
We have continued to expand out use of Google Big Query over the years. I'd say its flexibility and scalability is actually quite good. It also integrates well with other tools like Tableau and Power BI. It has served the needs of multiple data sources across multiple departments within my company.
Google Support has kindly provide individual support and consultants to assist with the integration work. In the circumstance where the consultants are not present to support with the work, Google Support Helpline will always be available to answer to the queries without having to wait for more than 3 days.
Previously, running complex queries on our on-premise data warehouse could take hours. Google BigQuery processes the same queries in minutes. We estimate it saves our team at least 25% of their time.
We can target our marketing campaigns very easily and understand our customer behaviour. It lets us personalize marketing campaigns and product recommendations and experience at least a 20% improvement in overall campaign performance.
Now, we only pay for the resources we use. Saved $1 million annually on data infrastructure and data storage costs compared to our previous solution.