The Aerospike Real-time Data Platform aims to enable organizations to act instantly across billions of transactions while reducing server footprint up to 80%. The vendor states Aerospike multi-cloud platform powers real-time applications with predictable sub-millisecond performance up to petabyte scale with five-nines uptime with globally distributed, consistent data. Aerospike boasts customers such as Airtel, Experian, European Central Bank, Nielsen, PayPal, Snap, Verizon Media and Wayfair.
N/A
Amazon DynamoDB
Score 8.7 out of 10
N/A
Amazon DynamoDB is a cloud-native, NoSQL, serverless database service.
Compared to the above for K/V lookups and writes, it is faster. However, less than 1 MB, i'd use Redis, if you're willing to write package for HA in Redis. However HA between Redis and aerospike, aerospike is top notch. K/V lookups were 20-30% faster than Redis, 50% faster …
We were developing an advertisement time auction application, where we had to store the client's personal details, advertisement-related details, location, and many other details. Moreover, we required a promotion, cookies, and a few more details from the front end. All this information is heavy in terms of size and cannot be lost if the server crash. So, we required an extremely fast disk database with high scalability and low throughput.
It’s great for server less and real-time applications. It would be great for gaming and mobile apps. However, if you need relational database and have fixed budget, do not use it. While budget can be managed, you need to be careful. Also this is not a tool for storing big data, there are other wide-column database types you could use for it ins the ad
If money isn't an issue, and you're not on the cloud, then I'd go with Aerospike. If you're the cloud ie, aws or azure, then i'd stick with dynamoDB or Cosmos then. Aerospike is definitely not something you want to put into the cloud. It doesn't work well w/ cross regions. If cross DC, you'll have to write some stuff for data integrity checks.
It's core to our business, we couldn't survive without it. We use it to drive everything from FTP logins to processing stories and delivering them to clients. It's reliable and easy to query from all of our pipeline services. Integration with things like AWS Lambda makes it easy to trigger events and run code whenever something changes in the database.
Functionally, DynamoDB has the features needed to use it. The interface is not as easy to use, which impacts its usability. Being familiar with AWS in general is helpful in understanding the interface, however it would be better if the interface more closely aligned with traditional tools for managing datastores.
It works very well across all the regions and response time is also very quick due to AWS's internal data transfer. Plus if your product requires HIPPA or some other regulations needs to be followed, you can easily replicate the DB into multiple regions and they manage all by it's own.
The only thing that can be compared to DynamoDB from the selected services can be Aurora. It is just that we use Aurora for High-Performance requirements as it can be 6 times faster than normal RDS DB. Both of them have served as well in the required scenario and we are very happy with most of the AWS services.
I have taken one point away due to its size limits. In case the application requires queries, it becomes really complicated to read and write data. When it comes to extremely large data sets such as the case in my company, a third-party logistics company, where huge amount of data is generated on a daily basis, even though the scalability is good, it becomes difficult to manage all the data due to limits.
Some developers see DynamoDB and try to fit problems to it, instead of picking the best solution for a given problem. This is true of any newer tool that people are trying to adopt.
It has allowed us to add more scalability to some of our systems.
As with any new technology there was a ramp up/rework phase as we learned best practices.