How to best handle data stored in different locations in Google BigQuery?

前端 未结 2 348
清歌不尽
清歌不尽 2021-01-13 01:22

My current workflow in BigQuery is as follows:

(1) query data in a public repository (stored in the US), (2) write it to a table in my repository, (3) export a csv

2条回答
  •  轻奢々
    轻奢々 (楼主)
    2021-01-13 01:57

    No matter what, you have data in the US that you need in the EU, so I think you have two options:

    1. You could continue to pay many smaller fees to move your reduced datasets from the US to the EU as you're doing today.

    2. You could pay the one-off fee to transfer the original public BQ dataset from the US to your own dataset in the EU. From then on, all queries you run stay in the same region, and you have no more trans-continental transfers.

    It really depends on how many queries you plan to do. If it's not a lot, then the way you're doing things today seems like it'd be the most efficient. If it's a lot, then moving the data once (paying the up-front fee) might work out cheaper.

    Maybe Google has some magical way to make this better, but as far as I can tell, you're dealing with lots of data on one side of the Atlantic that you need on the other side, and moving it across that wire costs money.

提交回复
热议问题