com.google.cloud.bigquery.BigQuery Maven / Gradle / Ivy
/*
* Copyright 2015 Google LLC
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.google.cloud.bigquery;
import static com.google.common.base.Preconditions.checkArgument;
import com.google.api.core.BetaApi;
import com.google.api.core.InternalApi;
import com.google.api.gax.paging.Page;
import com.google.cloud.FieldSelector;
import com.google.cloud.FieldSelector.Helper;
import com.google.cloud.Policy;
import com.google.cloud.RetryOption;
import com.google.cloud.Service;
import com.google.cloud.bigquery.spi.v2.BigQueryRpc;
import com.google.common.base.Function;
import com.google.common.collect.ImmutableList;
import com.google.common.collect.Lists;
import java.io.Serializable;
import java.util.ArrayList;
import java.util.List;
import org.checkerframework.checker.nullness.qual.NonNull;
/**
* An interface for Google Cloud BigQuery.
*
* @see Google Cloud BigQuery
*/
public interface BigQuery extends Service {
/**
* Fields of a BigQuery Dataset resource.
*
* @see Dataset
* Resource
*/
enum DatasetField implements FieldSelector {
ACCESS("access"),
CREATION_TIME("creationTime"),
DATASET_REFERENCE("datasetReference"),
DEFAULT_TABLE_EXPIRATION_MS("defaultTableExpirationMsS"),
DESCRIPTION("description"),
ETAG("etag"),
FRIENDLY_NAME("friendlyName"),
ID("id"),
LABELS("labels"),
LAST_MODIFIED_TIME("lastModifiedTime"),
LOCATION("location"),
SELF_LINK("selfLink");
static final List extends FieldSelector> REQUIRED_FIELDS =
ImmutableList.of(DATASET_REFERENCE);
private final String selector;
DatasetField(String selector) {
this.selector = selector;
}
@Override
public String getSelector() {
return selector;
}
}
/**
* Fields of a BigQuery Table resource.
*
* @see Table
* Resource
*/
enum TableField implements FieldSelector {
CREATION_TIME("creationTime"),
DESCRIPTION("description"),
ETAG("etag"),
EXPIRATION_TIME("expirationTime"),
EXTERNAL_DATA_CONFIGURATION("externalDataConfiguration"),
FRIENDLY_NAME("friendlyName"),
ID("id"),
LABELS("labels"),
LAST_MODIFIED_TIME("lastModifiedTime"),
LOCATION("location"),
NUM_BYTES("numBytes"),
NUM_LONG_TERM_BYTES("numLongTermBytes"),
NUM_ROWS("numRows"),
SCHEMA("schema"),
SELF_LINK("selfLink"),
STREAMING_BUFFER("streamingBuffer"),
TABLE_REFERENCE("tableReference"),
TIME_PARTITIONING("timePartitioning"),
RANGE_PARTITIONING("rangePartitioning"),
TYPE("type"),
VIEW("view");
static final List extends FieldSelector> REQUIRED_FIELDS =
ImmutableList.of(TABLE_REFERENCE, TYPE);
private final String selector;
TableField(String selector) {
this.selector = selector;
}
@Override
public String getSelector() {
return selector;
}
}
/**
* Metadata of a BigQuery Table.
*
* @see Table
* Resource
*/
enum TableMetadataView {
BASIC,
FULL,
STORAGE_STATS,
TABLE_METADATA_VIEW_UNSPECIFIED;
}
/**
* Fields of a BigQuery Model resource.
*
* @see Model
* Resource
*/
enum ModelField implements FieldSelector {
CREATION_TIME("creationTime"),
DESCRIPTION("description"),
ETAG("etag"),
EXPIRATION_TIME("expirationTime"),
FRIENDLY_NAME("friendlyName"),
LABELS("labels"),
LAST_MODIFIED_TIME("lastModifiedTime"),
LOCATION("location"),
MODEL_REFERENCE("modelReference"),
TRAINING_RUNS("trainingRuns"),
LABEL_COLUMNS("labelColumns"),
FEATURE_COLUMNS("featureColumns"),
TYPE("modelType");
static final List extends FieldSelector> REQUIRED_FIELDS = ImmutableList.of(MODEL_REFERENCE);
private final String selector;
ModelField(String selector) {
this.selector = selector;
}
@Override
public String getSelector() {
return selector;
}
}
/**
* Fields of a BigQuery Routine resource.
*
* @see Routine
* Resource
*/
enum RoutineField implements FieldSelector {
ARGUMENTS("arguments"),
CREATION_TIME("creationTime"),
DEFINITION_BODY("definitionBody"),
ETAG("etag"),
IMPORTED_LIBRARIES("importedLibraries"),
LANGUAGE("language"),
LAST_MODIFIED_TIME("lastModifiedTime"),
RETURN_TYPE("returnType"),
ROUTINE_REFERENCE("routineReference"),
ROUTINE_TYPE("routineType");
static final List extends FieldSelector> REQUIRED_FIELDS =
ImmutableList.of(ROUTINE_REFERENCE);
private final String selector;
RoutineField(String selector) {
this.selector = selector;
}
@Override
public String getSelector() {
return selector;
}
}
/**
* Fields of a BigQuery Job resource.
*
* @see Job Resource
*
*/
enum JobField implements FieldSelector {
CONFIGURATION("configuration"),
ETAG("etag"),
ID("id"),
JOB_REFERENCE("jobReference"),
SELF_LINK("selfLink"),
STATISTICS("statistics"),
STATUS("status"),
USER_EMAIL("user_email");
static final List extends FieldSelector> REQUIRED_FIELDS =
ImmutableList.of(JOB_REFERENCE, CONFIGURATION);
private final String selector;
JobField(String selector) {
this.selector = selector;
}
@Override
public String getSelector() {
return selector;
}
}
/** Class for specifying dataset list options. */
class DatasetListOption extends Option {
private static final long serialVersionUID = 8660294969063340498L;
private DatasetListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify a label filter. See
* https://cloud.google.com/bigquery/docs/adding-using-labels#filtering_datasets_using_labels
*
* @param labelFilter In the form "labels.key:value"
*/
public static DatasetListOption labelFilter(String labelFilter) {
return new DatasetListOption(BigQueryRpc.Option.LABEL_FILTER, labelFilter);
}
/** Returns an option to specify the maximum number of datasets returned per page. */
public static DatasetListOption pageSize(long pageSize) {
return new DatasetListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing datasets. */
public static DatasetListOption pageToken(String pageToken) {
return new DatasetListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
/** Returns an options to list all datasets, even hidden ones. */
public static DatasetListOption all() {
return new DatasetListOption(BigQueryRpc.Option.ALL_DATASETS, true);
}
}
/** Class for specifying dataset get, create and update options. */
class DatasetOption extends Option {
private static final long serialVersionUID = 1674133909259913250L;
private DatasetOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify the dataset's fields to be returned by the RPC call. If this
* option is not provided all dataset's fields are returned. { code DatasetOption.fields} can be
* used to specify only the fields of interest. {@link Dataset#getDatasetId()} is always
* returned, even if not specified.
*/
public static DatasetOption fields(DatasetField... fields) {
return new DatasetOption(
BigQueryRpc.Option.FIELDS, Helper.selector(DatasetField.REQUIRED_FIELDS, fields));
}
}
/** Class for specifying dataset delete options. */
class DatasetDeleteOption extends Option {
private static final long serialVersionUID = -7166083569900951337L;
private DatasetDeleteOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to delete a dataset even if non-empty. If not provided, attempting to
* delete a non-empty dataset will result in a {@link BigQueryException} being thrown.
*/
public static DatasetDeleteOption deleteContents() {
return new DatasetDeleteOption(BigQueryRpc.Option.DELETE_CONTENTS, true);
}
}
/** Class for specifying table list options. */
class ModelListOption extends Option {
private static final long serialVersionUID = 8660294969063322498L;
private ModelListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to specify the maximum number of models returned per page. */
public static ModelListOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new ModelListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing models. */
public static ModelListOption pageToken(String pageToken) {
return new ModelListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
}
/** Class for specifying routine list options. */
class RoutineListOption extends Option {
private static final long serialVersionUID = 8660294969063312498L;
private RoutineListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to specify the maximum number of routines returned per page. */
public static RoutineListOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new RoutineListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing routines. */
public static RoutineListOption pageToken(String pageToken) {
return new RoutineListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
}
/** Class for specifying table list options. */
class TableListOption extends Option {
private static final long serialVersionUID = 8660294969063340498L;
private TableListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to specify the maximum number of tables returned per page. */
public static TableListOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new TableListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing tables. */
public static TableListOption pageToken(String pageToken) {
return new TableListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
}
/** Class for specifying table get, create and update options. */
class TableOption extends Option {
private static final long serialVersionUID = -1723870134095936772L;
private TableOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify the table's fields to be returned by the RPC call. If this
* option is not provided all table's fields are returned. {@code TableOption.fields} can be
* used to specify only the fields of interest. {@link Table#getTableId()} and type (which is
* part of {@link Table#getDefinition()}) are always returned, even if not specified.
*/
public static TableOption fields(TableField... fields) {
return new TableOption(
BigQueryRpc.Option.FIELDS, Helper.selector(TableField.REQUIRED_FIELDS, fields));
}
/**
* Returns an option to specify the schema of the table (only applicable for external tables)
* should be autodetected when updating the table from the underlying source.
*/
public static TableOption autodetectSchema(boolean autodetect) {
return new TableOption(BigQueryRpc.Option.AUTODETECT_SCHEMA, autodetect);
}
/** Returns an option to specify the metadata of the table. */
public static TableOption tableMetadataView(TableMetadataView tableMetadataView) {
return new TableOption(BigQueryRpc.Option.TABLE_METADATA_VIEW, tableMetadataView);
}
}
/* Class for specifying IAM options. */
class IAMOption extends Option {
private static final long serialVersionUID = 8607992885371024269L;
private IAMOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
public static IAMOption requestedPolicyVersion(long version) {
return new IAMOption(BigQueryRpc.Option.REQUESTED_POLICY_VERSION, version);
}
}
/** Class for specifying model get, create and update options. */
class ModelOption extends Option {
private static final long serialVersionUID = -1723870134095226772L;
private ModelOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify the model's fields to be returned by the RPC call. If this
* option is not provided all model's fields are returned. {@code ModelOption.fields} can be
* used to specify only the fields of interest.
*/
public static ModelOption fields(ModelField... fields) {
return new ModelOption(
BigQueryRpc.Option.FIELDS, Helper.selector(ModelField.REQUIRED_FIELDS, fields));
}
}
/** Class for specifying table get, create and update options. */
class RoutineOption extends Option {
private static final long serialVersionUID = -1723870122095226772L;
private RoutineOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify the routines's fields to be returned by the RPC call. If this
* option is not provided all model's fields are returned. {@code RoutineOption.fields} can be
* used to specify only the fields of interest.
*/
public static RoutineOption fields(RoutineField... fields) {
return new RoutineOption(
BigQueryRpc.Option.FIELDS, Helper.selector(RoutineField.REQUIRED_FIELDS, fields));
}
}
/** Class for specifying table data list options. */
class TableDataListOption extends Option {
private static final long serialVersionUID = 8488823381738864434L;
private TableDataListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to specify the maximum number of rows returned per page. */
public static TableDataListOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new TableDataListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing table data. */
public static TableDataListOption pageToken(String pageToken) {
return new TableDataListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
/**
* Returns an option that sets the zero-based index of the row from which to start listing table
* data.
*/
public static TableDataListOption startIndex(long index) {
checkArgument(index >= 0);
return new TableDataListOption(BigQueryRpc.Option.START_INDEX, index);
}
}
/** Class for specifying job list options. */
class JobListOption extends Option {
private static final long serialVersionUID = -8207122131226481423L;
private JobListOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to list all jobs, even the ones issued by other users. */
public static JobListOption allUsers() {
return new JobListOption(BigQueryRpc.Option.ALL_USERS, true);
}
/** Returns an option to list only jobs that match the provided state filters. */
public static JobListOption stateFilter(JobStatus.State... stateFilters) {
List stringFilters =
Lists.transform(
ImmutableList.copyOf(stateFilters),
new Function() {
@Override
public String apply(JobStatus.State state) {
return state.name().toLowerCase();
}
});
return new JobListOption(BigQueryRpc.Option.STATE_FILTER, stringFilters);
}
/** Returns an option to filter out jobs before the given minimum creation time. */
public static JobListOption minCreationTime(long minCreationTime) {
return new JobListOption(BigQueryRpc.Option.MIN_CREATION_TIME, minCreationTime);
}
/** Returns an option to filter out jobs after the given maximum creation time. */
public static JobListOption maxCreationTime(long maxCreationTime) {
return new JobListOption(BigQueryRpc.Option.MAX_CREATION_TIME, maxCreationTime);
}
/** Returns an option to specify the maximum number of jobs returned per page. */
public static JobListOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new JobListOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start listing jobs. */
public static JobListOption pageToken(String pageToken) {
return new JobListOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
/** Returns an option to list only child job from specify parent job id. */
public static JobListOption parentJobId(String parentJobId) {
return new JobListOption(BigQueryRpc.Option.PARENT_JOB_ID, parentJobId);
}
/**
* Returns an option to specify the job's fields to be returned by the RPC call. If this option
* is not provided all job's fields are returned. {@code JobOption.fields()} can be used to
* specify only the fields of interest. {@link Job#getJobId()}, {@link JobStatus#getState()},
* {@link JobStatus#getError()} as well as type-specific configuration (e.g. {@link
* QueryJobConfiguration#getQuery()} for Query Jobs) are always returned, even if not specified.
* {@link JobField#SELF_LINK} and {@link JobField#ETAG} can not be selected when listing jobs.
*/
public static JobListOption fields(JobField... fields) {
return new JobListOption(
BigQueryRpc.Option.FIELDS,
Helper.listSelector("jobs", JobField.REQUIRED_FIELDS, fields, "state", "errorResult"));
}
}
/** Class for specifying table get and create options. */
class JobOption extends Option {
private static final long serialVersionUID = -3111736712316353665L;
private JobOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/**
* Returns an option to specify the job's fields to be returned by the RPC call. If this option
* is not provided all job's fields are returned. {@code JobOption.fields()} can be used to
* specify only the fields of interest. {@link Job#getJobId()} as well as type-specific
* configuration (e.g. {@link QueryJobConfiguration#getQuery()} for Query Jobs) are always
* returned, even if not specified.
*/
public static JobOption fields(JobField... fields) {
return new JobOption(
BigQueryRpc.Option.FIELDS, Helper.selector(JobField.REQUIRED_FIELDS, fields));
}
}
/** Class for specifying query results options. */
class QueryResultsOption extends Option {
private static final long serialVersionUID = 3788898503226985525L;
private QueryResultsOption(BigQueryRpc.Option option, Object value) {
super(option, value);
}
/** Returns an option to specify the maximum number of rows returned per page. */
public static QueryResultsOption pageSize(long pageSize) {
checkArgument(pageSize >= 0);
return new QueryResultsOption(BigQueryRpc.Option.MAX_RESULTS, pageSize);
}
/** Returns an option to specify the page token from which to start getting query results. */
public static QueryResultsOption pageToken(String pageToken) {
return new QueryResultsOption(BigQueryRpc.Option.PAGE_TOKEN, pageToken);
}
/**
* Returns an option that sets the zero-based index of the row from which to start getting query
* results.
*/
public static QueryResultsOption startIndex(long startIndex) {
checkArgument(startIndex >= 0);
return new QueryResultsOption(BigQueryRpc.Option.START_INDEX, startIndex);
}
/**
* Returns an option that sets how long to wait for the query to complete, in milliseconds,
* before returning. Default is 10 seconds.
*/
public static QueryResultsOption maxWaitTime(long maxWaitTime) {
checkArgument(maxWaitTime >= 0);
return new QueryResultsOption(BigQueryRpc.Option.TIMEOUT, maxWaitTime);
}
}
class QueryOption implements Serializable {
private static final long serialVersionUID = 6206193419355824689L;
private final Object option;
private QueryOption(Object option) {
this.option = option;
}
public QueryResultsOption getQueryResultsOption() {
return option instanceof QueryResultsOption ? (QueryResultsOption) option : null;
}
public RetryOption getRetryOption() {
return option instanceof RetryOption ? (RetryOption) option : null;
}
static QueryResultsOption[] filterQueryResultsOptions(QueryOption... options) {
List queryResultOptions = new ArrayList<>(options.length);
for (QueryOption opt : options) {
if (opt.getQueryResultsOption() != null) {
queryResultOptions.add(opt.getQueryResultsOption());
}
}
return queryResultOptions.toArray(new QueryResultsOption[queryResultOptions.size()]);
}
static RetryOption[] filterRetryOptions(QueryOption... options) {
List retryOptions = new ArrayList<>(options.length);
for (QueryOption opt : options) {
if (opt.getRetryOption() != null) {
retryOptions.add(opt.getRetryOption());
}
}
return retryOptions.toArray(new RetryOption[retryOptions.size()]);
}
public static QueryOption of(QueryResultsOption resultsOption) {
return new QueryOption(resultsOption);
}
public static QueryOption of(RetryOption waitOption) {
return new QueryOption(waitOption);
}
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
QueryOption that = (QueryOption) o;
return option != null ? option.equals(that.option) : that.option == null;
}
@Override
public int hashCode() {
return option != null ? option.hashCode() : 0;
}
}
/**
* Creates a new dataset.
*
* Example of creating a dataset.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* Dataset dataset = null;
* DatasetInfo datasetInfo = DatasetInfo.newBuilder(datasetName).build();
* try {
* // the dataset was created
* dataset = bigquery.create(datasetInfo);
* } catch (BigQueryException e) {
* // the dataset was not created
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Dataset create(DatasetInfo datasetInfo, DatasetOption... options);
/**
* Creates a new table.
*
* Example of creating a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* String fieldName = "string_field";
* TableId tableId = TableId.of(datasetName, tableName);
* // Table field definition
* Field field = Field.of(fieldName, LegacySQLTypeName.STRING);
* // Table schema definition
* Schema schema = Schema.of(field);
* TableDefinition tableDefinition = StandardTableDefinition.of(schema);
* TableInfo tableInfo = TableInfo.newBuilder(tableId, tableDefinition).build();
* Table table = bigquery.create(tableInfo);
* }
*
*
* @throws BigQueryException upon failure
*/
Table create(TableInfo tableInfo, TableOption... options);
/**
* Creates a new routine.
*
* @throws BigQueryException upon failure
*/
Routine create(RoutineInfo routineInfo, RoutineOption... options);
/**
* Creates a new job.
*
* Example of loading a newline-delimited-json file with textual fields from GCS to a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* String sourceUri = "gs://cloud-samples-data/bigquery/us-states/us-states.json";
* TableId tableId = TableId.of(datasetName, tableName);
* // Table field definition
* Field[] fields = new Field[] { Field.of("name", LegacySQLTypeName.STRING),
* Field.of("post_abbr", LegacySQLTypeName.STRING) };
* // Table schema definition
* Schema schema = Schema.of(fields);
* LoadJobConfiguration configuration = LoadJobConfiguration.builder(tableId, sourceUri)
* .setFormatOptions(FormatOptions.json()).setCreateDisposition(CreateDisposition.CREATE_IF_NEEDED)
* .setSchema(schema).build();
* // Load the table
* Job loadJob = bigquery.create(JobInfo.of(configuration));
* loadJob = loadJob.waitFor();
* // Check the table
* System.out.println("State: " + loadJob.getStatus().getState());
* return ((StandardTableDefinition) bigquery.getTable(tableId).getDefinition()).getNumRows();
* }
*
*
* Example of creating a query job.
*
*
* {
* @code
* String query = "SELECT field FROM my_dataset_name.my_table_name";
* Job job = null;
* JobConfiguration jobConfiguration = QueryJobConfiguration.of(query);
* JobInfo jobInfo = JobInfo.of(jobConfiguration);
* try {
* job = bigquery.create(jobInfo);
* } catch (BigQueryException e) {
* // the job was not created
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Job create(JobInfo jobInfo, JobOption... options);
/**
* Creates a new BigQuery query connection used for executing queries (not the same as BigQuery
* connection properties). It uses the BigQuery Storage Read API for high throughput queries by
* default.
*
* Example of creating a query connection.
*
*
* {
* @code
* ConnectionSettings connectionSettings =
* ConnectionSettings.newBuilder()
* .setRequestTimeout(10L)
* .setMaxResults(100L)
* .setUseQueryCache(true)
* .build();
* Connection connection = bigquery.createConnection(connectionSettings);
* }
*
*
* @throws BigQueryException upon failure
* @param connectionSettings
*/
@BetaApi
Connection createConnection(@NonNull ConnectionSettings connectionSettings);
/**
* Creates a new BigQuery query connection used for executing queries (not the same as BigQuery
* connection properties). It uses the BigQuery Storage Read API for high throughput queries by
* default. This overloaded method creates a Connection with default ConnectionSettings for query
* execution where default values are set for numBufferedRows (20000), useReadApi (true),
* useLegacySql (false).
*
* Example of creating a query connection.
*
*
* {
* @code
* Connection connection = bigquery.createConnection();
* }
*
*
* @throws BigQueryException upon failure
*/
@BetaApi
Connection createConnection();
/**
* Returns the requested dataset or {@code null} if not found.
*
* Example of getting a dataset.
*
*
* {
* @code
* String datasetName = "my_dataset";
* Dataset dataset = bigquery.getDataset(datasetName);
* }
*
*
* @throws BigQueryException upon failure
*/
Dataset getDataset(String datasetId, DatasetOption... options);
/**
* Returns the requested dataset or {@code null} if not found.
*
* Example of getting a dataset.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* DatasetId datasetId = DatasetId.of(projectId, datasetName);
* Dataset dataset = bigquery.getDataset(datasetId);
* }
*
*
* @throws BigQueryException upon failure
*/
Dataset getDataset(DatasetId datasetId, DatasetOption... options);
/**
* Lists the project's datasets. This method returns partial information on each dataset: ({@link
* Dataset#getDatasetId()}, {@link Dataset#getFriendlyName()} and {@link
* Dataset#getGeneratedId()}). To get complete information use either {@link #getDataset(String,
* DatasetOption...)} or {@link #getDataset(DatasetId, DatasetOption...)}.
*
* Example of listing datasets, specifying the page size.
*
*
* {
* @code
* // List datasets in the default project
* Page<Dataset> datasets = bigquery.listDatasets(DatasetListOption.pageSize(100));
* for (Dataset dataset : datasets.iterateAll()) {
* // do something with the dataset
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Page listDatasets(DatasetListOption... options);
/**
* Lists the datasets in the provided project. This method returns partial information on each
* dataset: ({@link Dataset#getDatasetId()}, {@link Dataset#getFriendlyName()} and {@link
* Dataset#getGeneratedId()}). To get complete information use either {@link #getDataset(String,
* DatasetOption...)} or {@link #getDataset(DatasetId, DatasetOption...)}.
*
* Example of listing datasets in a project, specifying the page size.
*
*
* {
* @code
* String projectId = "my_project_id";
* // List datasets in a specified project
* Page<{@link Dataset}> datasets = bigquery.listDatasets(projectId, DatasetListOption.pageSize(100));
* for (Dataset dataset : datasets.iterateAll()) {
* // do something with the dataset
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Page listDatasets(String projectId, DatasetListOption... options);
/**
* Deletes the requested dataset.
*
* Example of deleting a dataset from its id, even if non-empty.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* boolean deleted = bigquery.delete(datasetName, DatasetDeleteOption.deleteContents());
* if (deleted) {
* // the dataset was deleted
* } else {
* // the dataset was not found
* }
* }
*
*
* @return {@code true} if dataset was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(String datasetId, DatasetDeleteOption... options);
/**
* Deletes the requested dataset.
*
* Example of deleting a dataset, even if non-empty.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* DatasetId datasetId = DatasetId.of(projectId, datasetName);
* boolean deleted = bigquery.delete(datasetId, DatasetDeleteOption.deleteContents());
* if (deleted) {
* // the dataset was deleted
* } else {
* // the dataset was not found
* }
* }
*
*
* @return {@code true} if dataset was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(DatasetId datasetId, DatasetDeleteOption... options);
/**
* Deletes the requested table.
*
* @deprecated Now that BigQuery datasets contain multiple resource types, this invocation is
* ambiguous. Please use more strongly typed version of {@code #delete} that leverages an
* non-ambiguous resource type identifier such as {@code TableId}.
* @return {@code true} if table was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
@Deprecated
boolean delete(String datasetId, String tableId);
/**
* Deletes the requested table.
*
* Example of deleting a table.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* TableId tableId = TableId.of(projectId, datasetName, tableName);
* boolean deleted = bigquery.delete(tableId);
* if (deleted) {
* // the table was deleted
* } else {
* // the table was not found
* }
* }
*
*
* @return {@code true} if table was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(TableId tableId);
/**
* Deletes the requested model.
*
* Example of deleting a model.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* String tableName = "my_model_name";
* ModelId modelId = ModelId.of(projectId, datasetName, modelName);
* boolean deleted = bigquery.delete(modelId);
* if (deleted) {
* // the model was deleted
* } else {
* // the model was not found
* }
* }
*
*
* @return {@code true} if model was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(ModelId modelId);
/**
* Deletes the requested routine.
*
* Example of deleting a routine.
*
*
{@code
* String projectId = "my_project_id";
* String datasetId = "my_dataset_id";
* String routineId = "my_routine_id";
* RoutineId routineId = RoutineId.of(projectId, datasetId, routineId);
* boolean deleted = bigquery.delete(routineId);
* if (deleted) {
* // the routine was deleted
* } else {
* // the routine was not found
* }
* }
*
* @return {@code true} if routine was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(RoutineId routineId);
/**
* Deletes the requested job.
*
* @return {@code true} if job was deleted, {@code false} if it was not found
* @throws BigQueryException upon failure
*/
boolean delete(JobId jobId);
/**
* Updates dataset information.
*
* Example of updating a dataset by changing its description.
*
*
*
* {
* @code
* // String datasetName = "my_dataset_name";
* // String tableName = "my_table_name";
* // String newDescription = "new_description";
*
* Table beforeTable = bigquery.getTable(datasetName, tableName);
* TableInfo tableInfo = beforeTable.toBuilder().setDescription(newDescription).build();
* Table afterTable = bigquery.update(tableInfo);
*
* }
*
*
*
*
* @throws BigQueryException upon failure
*/
Dataset update(DatasetInfo datasetInfo, DatasetOption... options);
/**
* Updates table information.
*
* Example of updating a table by changing its description.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* String newDescription = "new_description";
* Table beforeTable = bigquery.getTable(datasetName, tableName);
* TableInfo tableInfo = beforeTable.toBuilder().setDescription(newDescription).build();
* Table afterTable = bigquery.update(tableInfo);
* }
*
*
* Example of updating a table by changing its expiration.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* Table beforeTable = bigquery.getTable(datasetName, tableName);
*
* // Set table to expire 5 days from now.
* long expirationMillis = DateTime.now().plusDays(5).getMillis();
* TableInfo tableInfo = beforeTable.toBuilder().setExpirationTime(expirationMillis).build();
* Table afterTable = bigquery.update(tableInfo);
* }
*
*
* @throws BigQueryException upon failure
*/
Table update(TableInfo tableInfo, TableOption... options);
/**
* Updates model information.
*
* Example of updating a model by changing its description.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String modelName = "my_model_name";
* String newDescription = "new_description";
* Model beforeModel = bigquery.getModel(datasetName, modelName);
* ModelInfo modelInfo = beforeModel.toBuilder().setDescription(newDescription).build();
* Model afterModel = bigquery.update(modelInfo);
* }
*
*
* Example of updating a model by changing its expiration.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String modelName = "my_model_name";
* Model beforeModel = bigquery.getModel(datasetName, modelName);
*
* // Set model to expire 5 days from now.
* long expirationMillis = DateTime.now().plusDays(5).getMillis();
* ModelInfo modelInfo = beforeModel.toBuilder().setExpirationTime(expirationMillis).build();
* Model afterModel = bigquery.update(modelInfo);
* }
*
*
* @throws BigQueryException upon failure
*/
Model update(ModelInfo modelInfo, ModelOption... options);
/**
* Updates routine information.
*
* @throws BigQueryException upon failure
*/
Routine update(RoutineInfo routineInfo, RoutineOption... options);
/**
* Returns the requested table or {@code null} if not found.
*
* Example of getting a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* Table table = bigquery.getTable(datasetName, tableName);
* }
*
*
* @throws BigQueryException upon failure
*/
Table getTable(String datasetId, String tableId, TableOption... options);
/**
* Returns the requested table or {@code null} if not found.
*
* Example of getting a table.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* TableId tableId = TableId.of(projectId, datasetName, tableName);
* Table table = bigquery.getTable(tableId);
* }
*
*
* @throws BigQueryException upon failure
*/
Table getTable(TableId tableId, TableOption... options);
/**
* Returns the requested model or {@code null} if not found.
*
* @throws BigQueryException upon failure
*/
Model getModel(String datasetId, String modelId, ModelOption... options);
/**
* Returns the requested model or {@code null} if not found.
*
* Example of getting a model.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* String modelName = "my_model_name";
* ModelId modelId = ModelId.of(projectId, datasetName, tableName);
* Model model = bigquery.getModel(modelId);
* }
*
*
* @throws BigQueryException upon failure
*/
Model getModel(ModelId tableId, ModelOption... options);
/**
* Returns the requested routine or {@code null} if not found.
*
* @throws BigQueryException upon failure
*/
Routine getRoutine(String datasetId, String routineId, RoutineOption... options);
/**
* Returns the requested routine or {@code null} if not found.
*
* @throws BigQueryException upon failure
*/
Routine getRoutine(RoutineId routineId, RoutineOption... options);
/** Lists the routines in the specified dataset. */
Page listRoutines(String datasetId, RoutineListOption... options);
/** Lists the routines in the specified dataset. */
Page listRoutines(DatasetId datasetId, RoutineListOption... options);
/**
* Lists the tables in the dataset. This method returns partial information on each table: ({@link
* Table#getTableId()}, {@link Table#getFriendlyName()}, {@link Table#getGeneratedId()} and type,
* which is part of {@link Table#getDefinition()}). To get complete information use either {@link
* #getTable(TableId, TableOption...)} or {@link #getTable(String, String, TableOption...)}.
*
* Example of listing the tables in a dataset, specifying the page size.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* Page<Table> tables = bigquery.listTables(datasetName, TableListOption.pageSize(100));
* for (Table table : tables.iterateAll()) {
* // do something with the table
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Page listTables(String datasetId, TableListOption... options);
/**
* Lists the tables in the dataset. This method returns partial information on each table: ({@link
* Table#getTableId()}, {@link Table#getFriendlyName()}, {@link Table#getGeneratedId()} and type,
* which is part of {@link Table#getDefinition()}). To get complete information use either {@link
* #getTable(TableId, TableOption...)} or {@link #getTable(String, String, TableOption...)}.
*
* Example of listing the tables in a dataset.
*
*
* {
* @code
* String projectId = "my_project_id";
* String datasetName = "my_dataset_name";
* DatasetId datasetId = DatasetId.of(projectId, datasetName);
* Page<Table> tables = bigquery.listTables(datasetId, TableListOption.pageSize(100));
* for (Table table : tables.iterateAll()) {
* // do something with the table
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Page listTables(DatasetId datasetId, TableListOption... options);
/** Lists the models in the dataset. */
Page listModels(String datasetId, ModelListOption... options);
/** Lists the models in the dataset. */
Page listModels(DatasetId datasetId, ModelListOption... options);
/**
* @param tableId
* @return A list of the partition ids present in the partitioned table
*/
List listPartitions(TableId tableId);
/**
* Sends an insert all request.
*
* Example of inserting rows into a table without running a load job.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* TableId tableId = TableId.of(datasetName, tableName);
* // Values of the row to insert
* Map<String, Object> rowContent = new HashMap<>();
* rowContent.put("booleanField", true);
* // Bytes are passed in base64
* rowContent.put("bytesField", "Cg0NDg0="); // 0xA, 0xD, 0xD, 0xE, 0xD in base64
* // Records are passed as a map
* Map<String, Object> recordsContent = new HashMap<>();
* recordsContent.put("stringField", "Hello, World!");
* rowContent.put("recordField", recordsContent);
* InsertAllResponse response = bigquery.insertAll(InsertAllRequest.newBuilder(tableId).addRow("rowId", rowContent)
* // More rows can be added in the same RPC by invoking .addRow() on the
* // builder
* .build());
* if (response.hasErrors()) {
* // If any of the insertions failed, this lets you inspect the errors
* for (Entry<Long, List<BigQueryError>> entry : response.getInsertErrors().entrySet()) {
* // inspect row error
* }
* }
* }
*
*
* @throws BigQueryException upon failure
*/
InsertAllResponse insertAll(InsertAllRequest request);
/**
* Lists the table's rows.
*
* Example of listing table rows, specifying the page size.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* // This example reads the result 100 rows per RPC call. If there's no need
* // to limit the number,
* // simply omit the option.
* TableResult tableData = bigquery.listTableData(datasetName, tableName, TableDataListOption.pageSize(100));
* for (FieldValueList row : tableData.iterateAll()) {
* // do something with the row
* }
* }
*
*
* @throws BigQueryException upon failure
*/
TableResult listTableData(String datasetId, String tableId, TableDataListOption... options);
/**
* Lists the table's rows.
*
* Example of listing table rows, specifying the page size.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* TableId tableIdObject = TableId.of(datasetName, tableName);
* // This example reads the result 100 rows per RPC call. If there's no need
* // to limit the number,
* // simply omit the option.
* TableResult tableData = bigquery.listTableData(tableIdObject, TableDataListOption.pageSize(100));
* for (FieldValueList row : tableData.iterateAll()) {
* // do something with the row
* }
* }
*
*
* @throws BigQueryException upon failure
*/
TableResult listTableData(TableId tableId, TableDataListOption... options);
/**
* Lists the table's rows. If the {@code schema} is not {@code null}, it is available to the
* {@link FieldValueList} iterated over.
*
* Example of listing table rows with schema.
*
*
{@code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* Schema schema = ...;
* String field = "field";
* TableResult tableData = bigquery.listTableData(datasetName, tableName, schema);
* for (FieldValueList row : tableData.iterateAll()) {
* row.get(field);
* }
* }
*
* @throws BigQueryException upon failure
*/
TableResult listTableData(
String datasetId, String tableId, Schema schema, TableDataListOption... options);
/**
* Lists the table's rows. If the {@code schema} is not {@code null}, it is available to the
* {@link FieldValueList} iterated over.
*
* Example of listing table rows with schema.
*
*
* {
* @code
* Schema schema = Schema.of(Field.of("word", LegacySQLTypeName.STRING),
* Field.of("word_count", LegacySQLTypeName.STRING), Field.of("corpus", LegacySQLTypeName.STRING),
* Field.of("corpus_date", LegacySQLTypeName.STRING));
* TableResult tableData = bigquery.listTableData(TableId.of("bigquery-public-data", "samples", "shakespeare"),
* schema);
* FieldValueList row = tableData.getValues().iterator().next();
* System.out.println(row.get("word").getStringValue());
* }
*
*
* @throws BigQueryException upon failure
*/
TableResult listTableData(TableId tableId, Schema schema, TableDataListOption... options);
/**
* Returns the requested job or {@code null} if not found. If the location of the job is not "US"
* or "EU", {@link #getJob(JobId, JobOption...)} must be used instead.
*
* Example of getting a job.
*
*
* {
* @code
* String jobName = "my_job_name";
* Job job = bigquery.getJob(jobName);
* if (job == null) {
* // job was not found
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Job getJob(String jobId, JobOption... options);
/**
* Returns the requested job or {@code null} if not found. If the location of the job is not "US"
* or "EU", the {@code jobId} must specify the job location.
*
* Example of getting a job.
*
*
* {
* @code
* String jobName = "my_job_name";
* JobId jobIdObject = JobId.of(jobName);
* Job job = bigquery.getJob(jobIdObject);
* if (job == null) {
* // job was not found
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Job getJob(JobId jobId, JobOption... options);
/**
* Lists the jobs.
*
* Example of listing jobs, specifying the page size.
*
*
* {
* @code
* Page<Job> jobs = bigquery.listJobs(JobListOption.pageSize(100));
* for (Job job : jobs.iterateAll()) {
* // do something with the job
* }
* }
*
*
* @throws BigQueryException upon failure
*/
Page listJobs(JobListOption... options);
/**
* Sends a job cancel request. This call will return immediately. The job status can then be
* checked using either {@link #getJob(JobId, JobOption...)} or {@link #getJob(String,
* JobOption...)}).
*
* If the location of the job is not "US" or "EU", {@link #cancel(JobId)} must be used instead.
*
*
Example of cancelling a job.
*
*
* {
* @code
* String jobName = "my_job_name";
* boolean success = bigquery.cancel(jobName);
* if (success) {
* // job was cancelled
* } else {
* // job was not found
* }
* }
*
*
* @return {@code true} if cancel was requested successfully, {@code false} if the job was not
* found
* @throws BigQueryException upon failure
*/
boolean cancel(String jobId);
/**
* Sends a job cancel request. This call will return immediately. The job status can then be
* checked using either {@link #getJob(JobId, JobOption...)} or {@link #getJob(String,
* JobOption...)}).
*
* If the location of the job is not "US" or "EU", the {@code jobId} must specify the job
* location.
*
*
Example of cancelling a job.
*
*
* {
* @code
* String jobName = "my_job_name";
* JobId jobId = JobId.of(jobName);
* boolean success = bigquery.cancel(jobId);
* if (success) {
* // job was cancelled
* } else {
* // job was not found
* }
* }
*
*
* @return {@code true} if cancel was requested successfully, {@code false} if the job was not
* found
* @throws BigQueryException upon failure
*/
boolean cancel(JobId jobId);
/**
* Runs the query associated with the request, using an internally-generated random JobId.
*
* If the location of the job is not "US" or "EU", {@link #query(QueryJobConfiguration, JobId,
* JobOption...)} must be used instead.
*
*
This method cannot be used in conjuction with {@link QueryJobConfiguration#dryRun()}
* queries. Since dry-run queries are not actually executed, there's no way to retrieve results.
*
*
Example of running a query.
*
*
* {
* @code
* // BigQuery bigquery = BigQueryOptions.getDefaultInstance().getService();
* String query = "SELECT corpus FROM `bigquery-public-data.samples.shakespeare` GROUP BY corpus;";
* QueryJobConfiguration queryConfig = QueryJobConfiguration.newBuilder(query).build();
*
* // Print the results.
* for (FieldValueList row : bigquery.query(queryConfig).iterateAll()) {
* for (FieldValue val : row) {
* System.out.printf("%s,", val.toString());
* }
* System.out.printf("\n");
* }
* }
*
*
* This method supports query-related preview features via environmental variables (enabled by
* setting the {@code QUERY_PREVIEW_ENABLED} environment variable to "TRUE"). Specifically, this
* method supports:
*
*
* - Stateless queries: query execution without corresponding job metadata
*
*
* The behaviour of these preview features is controlled by the bigquery service as well
*
* @throws BigQueryException upon failure
* @throws InterruptedException if the current thread gets interrupted while waiting for the query
* to complete
* @throws JobException if the job completes unsuccessfully
*/
TableResult query(QueryJobConfiguration configuration, JobOption... options)
throws InterruptedException, JobException;
/**
* Runs the query associated with the request, using the given JobId.
*
* If the location of the job is not "US" or "EU", the {@code jobId} must specify the job
* location.
*
*
This method cannot be used in conjuction with {@link QueryJobConfiguration#dryRun()}
* queries. Since dry-run queries are not actually executed, there's no way to retrieve results.
*
*
See {@link #query(QueryJobConfiguration, JobOption...)} for examples on populating a {@link
* QueryJobConfiguration}.
*
* @throws BigQueryException upon failure
* @throws InterruptedException if the current thread gets interrupted while waiting for the query
* to complete
* @throws JobException if the job completes unsuccessfully
*/
TableResult query(QueryJobConfiguration configuration, JobId jobId, JobOption... options)
throws InterruptedException, JobException;
/**
* Returns results of the query associated with the provided job.
*
*
Users are encouraged to use {@link Job#getQueryResults(QueryResultsOption...)} instead.
*/
@InternalApi
QueryResponse getQueryResults(JobId jobId, QueryResultsOption... options);
/**
* Returns a channel to write data to be inserted into a BigQuery table. Data format and other
* options can be configured using the {@link WriteChannelConfiguration} parameter. If the job is
* not in "US" or "EU", {@link #writer(JobId, WriteChannelConfiguration)} must be used instead.
*
*
Example of creating a channel with which to write to a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* String csvData = "StringValue1\nStringValue2\n";
* TableId tableId = TableId.of(datasetName, tableName);
* WriteChannelConfiguration writeChannelConfiguration = WriteChannelConfiguration.newBuilder(tableId)
* .setFormatOptions(FormatOptions.csv()).build();
* TableDataWriteChannel writer = bigquery.writer(writeChannelConfiguration);
* // Write data to writer
* try {
* writer.write(ByteBuffer.wrap(csvData.getBytes(Charsets.UTF_8)));
* } finally {
* writer.close();
* }
* // Get load job
* Job job = writer.getJob();
* job = job.waitFor();
* LoadStatistics stats = job.getStatistics();
* return stats.getOutputRows();
* }
*
*
* Example of writing a local file to a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* Path csvPath = FileSystems.getDefault().getPath(".", "my-data.csv");
* String location = "us";
* TableId tableId = TableId.of(datasetName, tableName);
* WriteChannelConfiguration writeChannelConfiguration = WriteChannelConfiguration.newBuilder(tableId)
* .setFormatOptions(FormatOptions.csv()).build();
* // The location must be specified; other fields can be auto-detected.
* JobId jobId = JobId.newBuilder().setLocation(location).build();
* TableDataWriteChannel writer = bigquery.writer(jobId, writeChannelConfiguration);
* // Write data to writer
* try (OutputStream stream = Channels.newOutputStream(writer)) {
* Files.copy(csvPath, stream);
* }
* // Get load job
* Job job = writer.getJob();
* job = job.waitFor();
* LoadStatistics stats = job.getStatistics();
* return stats.getOutputRows();
* }
*
*
* @throws BigQueryException upon failure
*/
TableDataWriteChannel writer(WriteChannelConfiguration writeChannelConfiguration);
/**
* Returns a channel to write data to be inserted into a BigQuery table. Data format and other
* options can be configured using the {@link WriteChannelConfiguration} parameter. If the job is
* not in "US" or "EU", the {@code jobId} must contain the location of the job.
*
* Example of creating a channel with which to write to a table.
*
*
* {
* @code
* String datasetName = "my_dataset_name";
* String tableName = "my_table_name";
* String csvData = "StringValue1\nStringValue2\n";
* String location = "us";
* TableId tableId = TableId.of(datasetName, tableName);
* WriteChannelConfiguration writeChannelConfiguration = WriteChannelConfiguration.newBuilder(tableId)
* .setFormatOptions(FormatOptions.csv()).build();
* // The location must be specified; other fields can be auto-detected.
* JobId jobId = JobId.newBuilder().setLocation(location).build();
* TableDataWriteChannel writer = bigquery.writer(jobId, writeChannelConfiguration);
* // Write data to writer
* try {
* writer.write(ByteBuffer.wrap(csvData.getBytes(Charsets.UTF_8)));
* } finally {
* writer.close();
* }
* // Get load job
* Job job = writer.getJob();
* job = job.waitFor();
* LoadStatistics stats = job.getStatistics();
* return stats.getOutputRows();
* }
*
*/
TableDataWriteChannel writer(JobId jobId, WriteChannelConfiguration writeChannelConfiguration);
/** Gets the IAM policy for a specified table. */
Policy getIamPolicy(TableId tableId, IAMOption... options);
/** Sets the IAM policy for a specified table. */
Policy setIamPolicy(TableId tableId, Policy policy, IAMOption... options);
/**
* Tests whether the caller holds specific permissions on a BigQuery table. The returned list
* represents the subset of granted permissions.
*/
List testIamPermissions(TableId table, List permissions, IAMOption... options);
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy