MongoDB vs. Redis vs. Cassandra for a fast-write, temporary row storage solution

后端 未结 9 725
无人共我
无人共我 2021-01-29 20:38

I\'m building a system that tracks and verifies ad impressions and clicks. This means that there are a lot of insert commands (about 90/second average, peaking at 250) and some

相关标签:
9条回答
  • 2021-01-29 21:12

    I have hand-on experience with mongodb, couchdb and cassandra. I converted a lot of files to base64 string and insert these string into nosql.
    mongodb is the fastest. cassandra is slowest. couchdb is slow too.

    I think mysql would be much faster than all of them, but I didn't try mysql for my test case yet.

    0 讨论(0)
  • 2021-01-29 21:13

    Just found this: http://blog.axant.it/archives/236

    Quoting the most interesting part:

    This second graph is about Redis RPUSH vs Mongo $PUSH vs Mongo insert, and I find this graph to be really interesting. Up to 5000 entries mongodb $push is faster even when compared to Redis RPUSH, then it becames incredibly slow, probably the mongodb array type has linear insertion time and so it becomes slower and slower. mongodb might gain a bit of performances by exposing a constant time insertion list type, but even with the linear time array type (which can guarantee constant time look-up) it has its applications for small sets of data.

    I guess everything depends at least on data type and volume. Best advice probably would be to benchmark on your typical dataset and see yourself.

    0 讨论(0)
  • 2021-01-29 21:15

    All three solutions (four if you count flat-files) will give you blazing fast writes. The non-relational (nosql) solutions will give you tunable fault-tolerance as well for the purposes of disaster recovery.

    In terms of scale, our test environment, with only three MongoDB nodes, can handle 2-3k mixed transactions per second. At 8 nodes, we can handle 12k-15k mixed transactions per second. Cassandra can scale even higher. 250 reads is (or should be) no problem.

    The more important question is, what do you want to do with this data? Operational reporting? Time-series analysis? Ad-hoc pattern analysis? real-time reporting?

    MongoDB is a good option if you want the ability to do ad-hoc analysis based on multiple attributes within a collection. You can put up to 40 indexes on a collection, though the indexes will be stored in-memory, so watch for size. But the result is a flexible analytical solution.

    Cassandra is a key-value store. You define a static column or set of columns that will act as your primary index right up front. All queries run against Cassandra should be tuned to this index. You can put a secondary on it, but that's about as far as it goes. You can, of course, use MapReduce to scan the store for non-key attribution, but it will be just that: a serial scan through the store. Cassandra also doesn't have the notion of "like" or regex operations on the server nodes. If you want to find all customers where the first name starts with "Alex", you'll have to scan through the entire collection, pull the first name out for each entry and run it through a client-side regex.

    I'm not familiar enough with Redis to speak intelligently about it. Sorry.

    If you are evaluating non-relational platforms, you might also want to consider CouchDB and Riak.

    Hope this helps.

    0 讨论(0)
  • 2021-01-29 21:16

    If you have the choice (and need to move away from flat fies) I would go with Redis. Its blazingly fast, will comfortably handle the load you're talking about, but more importantly you won't have to manage the flushing/IO code. I understand its pretty straight forward but less code to manage is better than more.

    You will also get horizontal scaling options with Redis that you may not get with file based caching.

    0 讨论(0)
  • 2021-01-29 21:20

    I currently work for a very large ad network and we write to flat files :)

    I'm personally a Mongo fan, but frankly, Redis and Cassandra are unlikely to perform either better or worse. I mean, all you're doing is throwing stuff into memory and then flushing to disk in the background (both Mongo and Redis do this).

    If you're looking for blazing fast speed, the other option is to keep several impressions in local memory and then flush them disk every minute or so. Of course, this is basically what Mongo and Redis do for you. Not a real compelling reason to move.

    0 讨论(0)
  • 2021-01-29 21:25

    I can get around 30k inserts/sec with MongoDB on a simple $350 Dell. If you only need around 2k inserts/sec, I would stick with MongoDB and shard it for scalability. Maybe also look into doing something with Node.js or something similar to make things more asynchronous.

    0 讨论(0)
提交回复
热议问题