The faster method to move redis data to MySQL

后端 未结 1 893
后悔当初
后悔当初 2021-02-04 19:25

We have big shopping and product dealing system. We have faced lots problem with MySQL so after few r&D we planned to use Redis and we start integrating Redis in our system.

1条回答
  •  北海茫月
    2021-02-04 19:47

    Is their any other way to dump big data from Redis to MySQL?

    Redis has the possibility (using bgsave) to generate a dump of the data in a non blocking and consistent way.

    https://github.com/sripathikrishnan/redis-rdb-tools

    You could use Sripathi Krishnan's well-known package to parse a redis dump file (RDB) in Python, and populate the MySQL instance offline. Or you can convert the Redis dump to JSON format, and write scripts in any language you want to populate MySQL.

    This solution is only interesting if you want to copy the complete data of the Redis instance into MySQL.

    Does Redis have any trigger system that i can use to avoid the crons like queue system?

    Redis has no trigger concept, but nothing prevents you to post events in Redis queues each time something must be copied to MySQL. For instance, instead of:

    # Add an item to a user shopping cart
    RPUSH user::cart 
    

    you could execute:

    # Add an item to a user shopping cart
    MULTI
    RPUSH user::cart 
    RPUSH cart_to_mysql :
    EXEC
    

    The MULTI/EXEC block makes it atomic and consistent. Then you just have to write a little daemon waiting on items of the cart_to_mysql queue (using BLPOP commands). For each dequeued item, the daemon has to fetch the relevant data from Redis, and populate the MySQL instance.

    Redis fail our store data in file so is it possible to store that data directly to MySQL database?

    I'm not sure I understand the question here. But if you use the above solution, the latency between Redis updates and MySQL updates will be quite limited. So if Redis fails, you will only loose the very last operations (contrary to a solution based on cron jobs). It is of course not possible to have 100% consistency in the propagation of data though.

    0 讨论(0)
提交回复
热议问题