本文整理汇总了Java中com.google.cloud.bigquery.TableId类的典型用法代码示例。如果您正苦于以下问题:Java TableId类的具体用法?Java TableId怎么用?Java TableId使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
TableId类属于com.google.cloud.bigquery包,在下文中一共展示了TableId类的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。
示例1: deleteTableFromId
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of deleting a table.
*/
// [TARGET delete(TableId)]
// [VARIABLE "my_project_id"]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
public Boolean deleteTableFromId(String projectId, String datasetName, String tableName) {
// [START deleteTableFromId]
TableId tableId = TableId.of(projectId, datasetName, tableName);
Boolean deleted = bigquery.delete(tableId);
if (deleted) {
// the table was deleted
} else {
// the table was not found
}
// [END deleteTableFromId]
return deleted;
}
示例2: createTable
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of creating a table.
*/
// [TARGET create(TableInfo, TableOption...)]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
// [VARIABLE "string_field"]
public Table createTable(String datasetName, String tableName, Schema schema) {
// [START createTable]
TableId tableId = TableId.of(datasetName, tableName);
// Table field definition
//Field field = Field.of(fieldNames[0], Field.Type.string());
TableDefinition tableDefinition = StandardTableDefinition.of(schema);
TableInfo tableInfo = TableInfo.newBuilder(tableId, tableDefinition).build();
Table table = bigquery.create(tableInfo);
// [END createTable]
return table;
}
示例3: listTableDataFromId
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of listing table rows, specifying the page size.
*/
// [TARGET listTableData(TableId, TableDataListOption...)]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
public Page<List<FieldValue>> listTableDataFromId(String datasetName, String tableName) {
// [START listTableDataFromId]
TableId tableIdObject = TableId.of(datasetName, tableName);
Page<List<FieldValue>> tableData =
bigquery.listTableData(tableIdObject, TableDataListOption.pageSize(100));
Iterator<List<FieldValue>> rowIterator = tableData.iterateAll();
while (rowIterator.hasNext()) {
List<FieldValue> row = rowIterator.next();
// do something with the row
}
// [END listTableDataFromId]
return tableData;
}
示例4: testRetrieveSchema
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testRetrieveSchema() throws Exception {
final TableId table = TableId.of("test", "kafka_topic");
final String testTopic = "kafka-topic";
final String testSubject = "kafka-topic-value";
final String testAvroSchemaString =
"{\"type\": \"record\", "
+ "\"name\": \"testrecord\", "
+ "\"fields\": [{\"name\": \"f1\", \"type\": \"string\"}]}";
final SchemaMetadata testSchemaMetadata = new SchemaMetadata(1, 1, testAvroSchemaString);
SchemaRegistryClient schemaRegistryClient = mock(SchemaRegistryClient.class);
when(schemaRegistryClient.getLatestSchemaMetadata(testSubject)).thenReturn(testSchemaMetadata);
SchemaRegistrySchemaRetriever testSchemaRetriever = new SchemaRegistrySchemaRetriever(
schemaRegistryClient,
new AvroData(0)
);
Schema expectedKafkaConnectSchema =
SchemaBuilder.struct().field("f1", Schema.STRING_SCHEMA).name("testrecord").build();
assertEquals(expectedKafkaConnectSchema, testSchemaRetriever.retrieveSchema(table, testTopic));
}
示例5: getRecordTable
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
private PartitionedTableId getRecordTable(SinkRecord record) {
TableId baseTableId = topicsToBaseTableIds.get(record.topic());
PartitionedTableId.Builder builder = new PartitionedTableId.Builder(baseTableId);
if (useMessageTimeDatePartitioning) {
if (record.timestampType() == TimestampType.NO_TIMESTAMP_TYPE) {
throw new ConnectException("Message has no timestamp type, cannot use message timestamp to partition.");
}
builder.setDayPartition(record.timestamp());
} else {
builder.setDayPartitionForNow();
}
return builder.build();
}
示例6: testRetrieveSchemaWithMultipleSchemasSucceeds
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testRetrieveSchemaWithMultipleSchemasSucceeds() {
final String floatSchemaTopic = "test-float32";
final String intSchemaTopic = "test-int32";
final TableId floatTableId = getTableId("testFloatTable", "testFloatDataset");
final TableId intTableId = getTableId("testIntTable", "testIntDataset");
SchemaRetriever retriever = new MemorySchemaRetriever();
retriever.configure(new HashMap<>());
Schema expectedIntSchema = Schema.INT32_SCHEMA;
Schema expectedFloatSchema = Schema.OPTIONAL_FLOAT32_SCHEMA;
retriever.setLastSeenSchema(floatTableId, floatSchemaTopic, expectedFloatSchema);
retriever.setLastSeenSchema(intTableId, intSchemaTopic, expectedIntSchema);
Assert.assertEquals(retriever.retrieveSchema(floatTableId, floatSchemaTopic), expectedFloatSchema);
Assert.assertEquals(retriever.retrieveSchema(intTableId, intSchemaTopic), expectedIntSchema);
}
示例7: testTableIdBuilder
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testTableIdBuilder() {
final String project = "project";
final String dataset = "dataset";
final String table = "table";
final TableId tableId = TableId.of(project, dataset, table);
final PartitionedTableId partitionedTableId = new PartitionedTableId.Builder(tableId).build();
Assert.assertEquals(project, partitionedTableId.getProject());
Assert.assertEquals(dataset, partitionedTableId.getDataset());
Assert.assertEquals(table, partitionedTableId.getBaseTableName());
Assert.assertEquals(table, partitionedTableId.getFullTableName());
Assert.assertEquals(tableId, partitionedTableId.getBaseTableId());
Assert.assertEquals(tableId, partitionedTableId.getFullTableId());
}
示例8: testWithPartition
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testWithPartition() {
final String dataset = "dataset";
final String table = "table";
final LocalDate partitionDate = LocalDate.of(2016, 9, 21);
final PartitionedTableId partitionedTableId =
new PartitionedTableId.Builder(dataset, table).setDayPartition(partitionDate).build();
final String expectedPartition = "20160921";
Assert.assertEquals(dataset, partitionedTableId.getDataset());
Assert.assertEquals(table, partitionedTableId.getBaseTableName());
Assert.assertEquals(table + "$" + expectedPartition, partitionedTableId.getFullTableName());
final TableId expectedBaseTableId = TableId.of(dataset, table);
final TableId expectedFullTableId = TableId.of(dataset, table + "$" + expectedPartition);
Assert.assertEquals(expectedBaseTableId, partitionedTableId.getBaseTableId());
Assert.assertEquals(expectedFullTableId, partitionedTableId.getFullTableId());
}
示例9: testWithEpochTimePartition
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testWithEpochTimePartition() {
final String dataset = "dataset";
final String table = "table";
final long utcTime = 1509007584334L;
final PartitionedTableId partitionedTableId =
new PartitionedTableId.Builder(dataset, table).setDayPartition(utcTime).build();
final String expectedPartition = "20171026";
Assert.assertEquals(dataset, partitionedTableId.getDataset());
Assert.assertEquals(table, partitionedTableId.getBaseTableName());
Assert.assertEquals(table + "$" + expectedPartition, partitionedTableId.getFullTableName());
final TableId expectedBaseTableId = TableId.of(dataset, table);
final TableId expectedFullTableId = TableId.of(dataset, table + "$" + expectedPartition);
Assert.assertEquals(expectedBaseTableId, partitionedTableId.getBaseTableId());
Assert.assertEquals(expectedFullTableId, partitionedTableId.getFullTableId());
}
示例10: testNonAutoCreateTablesFailure
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test(expected = BigQueryConnectException.class)
public void testNonAutoCreateTablesFailure() {
final String dataset = "scratch";
final String existingTableTopic = "topic-with-existing-table";
final String nonExistingTableTopic = "topic-without-existing-table";
final TableId existingTable = TableId.of(dataset, "topic_with_existing_table");
final TableId nonExistingTable = TableId.of(dataset, "topic_without_existing_table");
Map<String, String> properties = propertiesFactory.getProperties();
properties.put(BigQuerySinkConnectorConfig.TABLE_CREATE_CONFIG, "false");
properties.put(BigQuerySinkConfig.SANITIZE_TOPICS_CONFIG, "true");
properties.put(BigQuerySinkConfig.DATASETS_CONFIG, String.format(".*=%s", dataset));
properties.put(
BigQuerySinkConfig.TOPICS_CONFIG,
String.format("%s, %s", existingTableTopic, nonExistingTableTopic)
);
BigQuery bigQuery = mock(BigQuery.class);
Table fakeTable = mock(Table.class);
when(bigQuery.getTable(existingTable)).thenReturn(fakeTable);
when(bigQuery.getTable(nonExistingTable)).thenReturn(null);
BigQuerySinkConnector testConnector = new BigQuerySinkConnector(bigQuery);
testConnector.start(properties);
}
示例11: testSimplePutWhenSchemaRetrieverIsNotNull
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
@Test
public void testSimplePutWhenSchemaRetrieverIsNotNull() {
final String topic = "test-topic";
Map<String, String> properties = propertiesFactory.getProperties();
properties.put(BigQuerySinkConfig.TOPICS_CONFIG, topic);
properties.put(BigQuerySinkConfig.DATASETS_CONFIG, ".*=scratch");
BigQuery bigQuery = mock(BigQuery.class);
SinkTaskContext sinkTaskContext = mock(SinkTaskContext.class);
InsertAllResponse insertAllResponse = mock(InsertAllResponse.class);
when(bigQuery.insertAll(anyObject())).thenReturn(insertAllResponse);
when(insertAllResponse.hasErrors()).thenReturn(false);
SchemaRetriever schemaRetriever = mock(SchemaRetriever.class);
BigQuerySinkTask testTask = new BigQuerySinkTask(bigQuery, schemaRetriever);
testTask.initialize(sinkTaskContext);
testTask.start(properties);
testTask.put(Collections.singletonList(spoofSinkRecord(topic)));
testTask.flush(Collections.emptyMap());
verify(bigQuery, times(1)).insertAll(any(InsertAllRequest.class));
verify(schemaRetriever, times(1)).setLastSeenSchema(any(TableId.class), any(String.class), any(Schema.class));
}
示例12: runQueryPermanentTable
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
public static void runQueryPermanentTable(
String queryString,
String destinationDataset,
String destinationTable,
boolean allowLargeResults) throws TimeoutException, InterruptedException {
QueryJobConfiguration queryConfig =
QueryJobConfiguration.newBuilder(queryString)
// Save the results of the query to a permanent table. See:
// https://cloud.google.com/bigquery/docs/writing-results#permanent-table
.setDestinationTable(TableId.of(destinationDataset, destinationTable))
// Allow results larger than the maximum response size.
// If true, a destination table must be set. See:
// https://cloud.google.com/bigquery/docs/writing-results#large-results
.setAllowLargeResults(allowLargeResults)
.build();
runQuery(queryConfig);
}
示例13: getTableFromId
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of getting a table.
*/
// [TARGET getTable(TableId, TableOption...)]
// [VARIABLE "my_project_id"]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
public Table getTableFromId(String projectId, String datasetName, String tableName) {
// [START getTableFromId]
TableId tableId = TableId.of(projectId, datasetName, tableName);
Table table = bigquery.getTable(tableId);
// [END getTableFromId]
return table;
}
示例14: writeToTable
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of creating a channel with which to write to a table.
*/
// [TARGET writer(WriteChannelConfiguration)]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
// [VARIABLE "StringValue1\nStringValue2\n"]
public long writeToTable(String datasetName, String tableName, String csvData)
throws IOException, InterruptedException, TimeoutException {
// [START writeToTable]
TableId tableId = TableId.of(datasetName, tableName);
WriteChannelConfiguration writeChannelConfiguration =
WriteChannelConfiguration.newBuilder(tableId)
.setFormatOptions(FormatOptions.csv())
.build();
TableDataWriteChannel writer = bigquery.writer(writeChannelConfiguration);
// Write data to writer
try {
writer.write(ByteBuffer.wrap(csvData.getBytes(Charsets.UTF_8)));
} finally {
writer.close();
}
// Get load job
Job job = writer.getJob();
job = job.waitFor();
LoadStatistics stats = job.getStatistics();
return stats.getOutputRows();
// [END writeToTable]
}
示例15: writeFileToTable
import com.google.cloud.bigquery.TableId; //导入依赖的package包/类
/**
* Example of riting a local file to a table.
*/
// [TARGET writer(WriteChannelConfiguration)]
// [VARIABLE "my_dataset_name"]
// [VARIABLE "my_table_name"]
// [VARIABLE FileSystems.getDefault().getPath(".", "my-data.csv")]
public long writeFileToTable(String datasetName, String tableName, Path fileFullPath, FormatOptions formatOptions)
throws IOException, InterruptedException, TimeoutException {
// [START writeFileToTable]
TableId tableId = TableId.of(datasetName, tableName);
WriteChannelConfiguration writeChannelConfiguration =
WriteChannelConfiguration.newBuilder(tableId)
.setFormatOptions(formatOptions)
.build();
TableDataWriteChannel writer = bigquery.writer(writeChannelConfiguration);
// Write data to writer
try (OutputStream stream = Channels.newOutputStream(writer)) {
Files.copy(fileFullPath, stream);
}
// Get load job
Job job = writer.getJob();
job = job.waitFor();
if(job.getStatus().getExecutionErrors() != null){
String errors = "";
for(BigQueryError error : job.getStatus().getExecutionErrors()){
errors += "error: " + error.getMessage() + ", reason: " + error.getReason() + ", location: " + error.getLocation() + "; ";
}
throw new IOException(errors);
}
LoadStatistics stats = job.getStatistics();
return stats.getOutputRows();
// [END writeFileToTable]
}