How to efficiently import many large JSON files directly from S3 into MongoDB

前端 未结 2 1706
忘了有多久
忘了有多久 2021-02-06 10:25

I have compressed JSON files in S3 and I would like to set up MongoDB in EC2 to server json documents contained in these files. The compressed files are >100M and there are 1000

相关标签:
2条回答
  • 2021-02-06 10:59

    The alternative of using s3cmd is to use aws s3 which has fewer features but comes installed with AWS CLI.

    The command would look like:

    aws S3 cp <your s3 URI> - | \
    mongoimport \
    --db <dbName> \
    --collection <collectionName> \
    

    The - sends the file to stdout

    0 讨论(0)
  • 2021-02-06 11:00

    You don't need to store intermediate files, you can pipe the output of s3 file to stdout and you can get input to mongoimport from stdin.

    Your full command would look something like:

    s3cmd get s3://<yourFilename> - | mongoimport -d <dbName> -c <collectionName>
    

    note the - which says send the file to stdout rather than to a filename.

    0 讨论(0)
提交回复
热议问题