Bigtable-BigQuery Import via DataFlow: 2 questions on table partitioning and Timestamps
问题 I have a job in Dataflow importing data from Bigtable into Bigquery by using built-in Dataflow APIs for both. I have two questions: Question 1: If the source data is in one large table in Bigtable, how can I partition it into a set of sub- or smaller tables in BigQuery dynamically based on, say, the given Bigtable row-key known only at run-time? The Java code in Dataflow looks like this: p.apply(Read.from(CloudBigtableIO.read(config))) .apply(ParDo.of(new SomeDoFNonBTSourceData())) .apply