BigTable

How to connect to a running bigtable emulator from java

微笑、不失礼 提交于 2019-12-19 07:39:20
问题 I am trying to use the bigtable emulator from gcloud beta emulators. I launch the emulator, grab the hostname (localhost) and port (in this instance 8885) gcloud beta emulators bigtable start Executing: /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/platform/bigtable-emulator/cbtemulator --host=localhost --port=8885 I am trying to connect to the emulator from a java test client, here is what I provide: Configuration conf = BigtableConfiguration.configure(projectId, instanceId);

Tree structures in a nosql database

怎甘沉沦 提交于 2019-12-14 00:14:36
问题 I'm developing an application for Google App Engine which uses BigTable for its datastore. It's an application about writing a story collaboratively. It's a very simple hobby project that I'm working on just for fun. It's open source and you can see it here: http://story.multifarce.com/ The idea is that anyone can write a paragraph, which then needs to be validated by two other people. A story can also be branched at any paragraph, so that another version of the story can continue in another

Read And Modify on same transaction - Bigtable

爱⌒轻易说出口 提交于 2019-12-11 07:34:20
问题 I building a coupon system and I use bigtable. My schema has two column - Customer ID, Coupon code I would like to query the table to check if the customer already exists and if true to return the code and if it doesn't, modify the customer id cell with the id and to return back the code. I saw there is an option to do it in Bigtable ReadModifyWriteRow operator or with CheckAndMutateRow but I not found any references. 回答1: Google has API documentation for Bigtable and Python available here.

Can't connect to Bigtable from a Spring Boot application

情到浓时终转凉″ 提交于 2019-12-11 07:04:20
问题 I have a standalone application that works fine with Bigtable when creating a connection like this: Connection connection = BigtableConfiguration.connect(PROJECT_ID, INSTANCE_ID) and using the following dependencies: <dependency> <groupId>com.google.apis</groupId> <artifactId>google-api-services-storage</artifactId> <version>v1-rev78-1.22.0</version> </dependency> <dependency> <groupId>com.google.apis</groupId> <artifactId>google-api-services-pubsub</artifactId> <version>v1-rev11-1.22.0<

Can Bloom Filters in BigTable be used to filter based only on row ID?

落花浮王杯 提交于 2019-12-11 03:32:37
问题 BigTable uses Bloom filters to allow point reads to avoid accessing SSTables that do not contain any data within a given key-column pair. Can these Bloom filters also be used to avoid accessing SSTables if the query only specifies the row ID and no column ID? BigTable uses row-column pairs as keys to insert into its bloom filters. This means that a query can use these filters for a point read that specifies a row-column pair. Now, suppose we have a query to get all columns of a row based only

Error exporting data from Google Cloud Bigtable

对着背影说爱祢 提交于 2019-12-11 00:23:26
问题 While going through the Google docs, I'm getting the below stack trace on the final export command (executed from the master instance with appropriate env variables set). ${HADOOP_HOME}/bin/hadoop jar ${HADOOP_BIGTABLE_JAR} export-table -libjars ${HADOOP_BIGTABLE_JAR} <table-name> <gs://bucket> SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/home/hadoop/hbase-install/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found

How to get filtered data from Bigtable using Python?

房东的猫 提交于 2019-12-10 10:33:57
问题 I am using Bigtable emulator and have successfully added a table in it and now I need to get filtered data. The table is as follows: arc_record_id | record_id | batch_id 1 |624 |86 2 |625 |86 3 |626 |86 and so on...till arc_record_id 10. I have tried this given below Python code: visit_dt_filter = ValueRangeFilter(start_value = "1".encode('utf-8'), end_value = "2".encode('utf-8')) col1_filter = ColumnQualifierRegexFilter(b'arc_record_id') chain1 = RowFilterChain(filters=[col1_filter, visit_dt

Bigtable CSV import

我怕爱的太早我们不能终老 提交于 2019-12-07 15:39:39
问题 I have a large csv dataset (>5TB) in multiple files (stored in a storage bucket) that I need to import into Google Bigtable. The files are in the format: rowkey,s1,s2,s3,s4 text,int,int,int,int ... There is an importtsv function with hbase that would be perfect but this does not seem to be available when using Google hbase shell in windows. Is it possible to use this tool? If not, what is the fastest way of achieving this? I have little experience with hbase and Google Cloud so a simple

Cannot connect from Titan to Google Bigtable via Hbase client

三世轮回 提交于 2019-12-07 14:12:00
问题 I am trying to connect to Titan 1.0.0 with Hadoop 2 (HBase 1.0.2 client) (available in https://github.com/thinkaurelius/titan/wiki/Downloads) with Google Cloud Bigtable service, using its HBase client. I could successfully connect to Bigtable from an HBase shell in a GCE instance, by following this procedure. The hbase-site.xml is according to the template, and I have downloaded Bigtable jars for ALPN_VERSION=8.1.5.v20150921 (Oracle JDK SE 1.8.0_60) <configuration> <property> <name>hbase

How to get filtered data from Bigtable using Python?

不羁岁月 提交于 2019-12-06 13:21:27
I am using Bigtable emulator and have successfully added a table in it and now I need to get filtered data. The table is as follows: arc_record_id | record_id | batch_id 1 |624 |86 2 |625 |86 3 |626 |86 and so on...till arc_record_id 10. I have tried this given below Python code: visit_dt_filter = ValueRangeFilter(start_value = "1".encode('utf-8'), end_value = "2".encode('utf-8')) col1_filter = ColumnQualifierRegexFilter(b'arc_record_id') chain1 = RowFilterChain(filters=[col1_filter, visit_dt_filter]) partial_rows = testTable.read_rows(filter_=chain1) for row in partial_rows: cell = row.cells