io.cdap.plugin.gcp.bigquery.sqlengine.BigQuerySelectDataset Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of google-cloud Show documentation
Show all versions of google-cloud Show documentation
Plugins for Google Big Query
The newest version!
package io.cdap.plugin.gcp.bigquery.sqlengine;
import com.google.cloud.bigquery.BigQuery;
import com.google.cloud.bigquery.Dataset;
import com.google.cloud.bigquery.DatasetId;
import com.google.cloud.bigquery.Job;
import com.google.cloud.bigquery.JobId;
import com.google.cloud.bigquery.JobInfo;
import com.google.cloud.bigquery.QueryJobConfiguration;
import com.google.cloud.bigquery.Table;
import com.google.cloud.bigquery.TableDefinition;
import com.google.cloud.bigquery.TableId;
import io.cdap.cdap.api.data.schema.Schema;
import io.cdap.cdap.api.metrics.Metrics;
import io.cdap.cdap.etl.api.engine.sql.SQLEngineException;
import io.cdap.cdap.etl.api.engine.sql.dataset.SQLDataset;
import io.cdap.plugin.gcp.bigquery.sink.BigQuerySinkUtils;
import io.cdap.plugin.gcp.bigquery.sqlengine.util.BigQuerySQLEngineUtils;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;
import javax.annotation.Nullable;
/**
* SQL Dataset that represents the result of a "Select" operation, such as join, that is executed in BigQuery.
*/
public class BigQuerySelectDataset implements SQLDataset, BigQuerySQLDataset {
private static final Logger LOG = LoggerFactory.getLogger(BigQuerySelectDataset.class);
private final String datasetName;
private final Schema outputSchema;
private final BigQuerySQLEngineConfig sqlEngineConfig;
private final BigQuery bigQuery;
private final String project;
private final DatasetId bqDataset;
private final String bqTable;
private final String jobId;
private final BigQueryJobType operation;
private final String selectQuery;
private final Metrics metrics;
private Long numRows;
public static BigQuerySelectDataset getInstance(String datasetName,
Schema outputSchema,
BigQuerySQLEngineConfig sqlEngineConfig,
BigQuery bigQuery,
String project,
DatasetId bqDataset,
String bqTable,
String jobId,
BigQueryJobType jobType,
String selectQuery,
Metrics metrics) {
return new BigQuerySelectDataset(datasetName,
outputSchema,
sqlEngineConfig,
bigQuery,
project,
bqDataset,
bqTable,
jobId,
jobType,
selectQuery,
metrics);
}
private BigQuerySelectDataset(String datasetName,
Schema outputSchema,
BigQuerySQLEngineConfig sqlEngineConfig,
BigQuery bigQuery,
String project,
DatasetId bqDataset,
String bqTable,
String jobId,
BigQueryJobType operation,
String selectQuery,
Metrics metrics) {
this.datasetName = datasetName;
this.outputSchema = outputSchema;
this.sqlEngineConfig = sqlEngineConfig;
this.bigQuery = bigQuery;
this.project = project;
this.bqDataset = bqDataset;
this.bqTable = bqTable;
this.jobId = jobId;
this.operation = operation;
this.selectQuery = selectQuery;
this.metrics = metrics;
}
public BigQuerySelectDataset execute() {
TableId destinationTable = TableId.of(bqDataset.getProject(), bqDataset.getDataset(), bqTable);
// Get location for target dataset. This way, the job will run in the same location as the dataset
Dataset dataset = bigQuery.getDataset(bqDataset);
String location = dataset.getLocation();
// Update destination table schema to match configured schema in the pipeline.
updateTableSchema(destinationTable, outputSchema);
LOG.info("Creating table `{}` using job: {} with SQL statement: {}", bqTable, jobId,
selectQuery);
// Run BigQuery job with supplied SQL statement, storing results in a new table
QueryJobConfiguration queryConfig =
QueryJobConfiguration.newBuilder(selectQuery)
.setDestinationTable(destinationTable)
.setCreateDisposition(JobInfo.CreateDisposition.CREATE_NEVER)
.setWriteDisposition(JobInfo.WriteDisposition.WRITE_APPEND)
.setSchemaUpdateOptions(Collections.singletonList(JobInfo.SchemaUpdateOption.ALLOW_FIELD_ADDITION))
.setPriority(sqlEngineConfig.getJobPriority())
.setLabels(BigQuerySQLEngineUtils.getJobTags(operation))
.build();
// Create a job ID so that we can safely retry.
JobId bqJobId = JobId.newBuilder().setJob(jobId).setLocation(location).setProject(project).build();
Job queryJob = bigQuery.create(JobInfo.newBuilder(queryConfig).setJobId(bqJobId).build());
// Wait for the query to complete.
try {
queryJob = queryJob.waitFor();
} catch (InterruptedException ie) {
throw new SQLEngineException("Interrupted exception when executing Join operation", ie);
}
// Check for errors
if (queryJob == null) {
throw new SQLEngineException("BigQuery job not found: " + jobId);
} else if (queryJob.getStatus().getError() != null) {
BigQuerySQLEngineUtils.logJobMetrics(queryJob, metrics);
throw new SQLEngineException(String.format(
"Error executing BigQuery Job: '%s' in Project '%s', Dataset '%s', Location'%s' : %s",
jobId, project, bqDataset, location, queryJob.getStatus().getError().toString()));
}
LOG.info("Created BigQuery table `{}` using Job: {}", bqTable, jobId);
BigQuerySQLEngineUtils.logJobMetrics(queryJob, metrics);
return this;
}
@Override
public String getDatasetName() {
return datasetName;
}
@Override
public Schema getSchema() {
return outputSchema;
}
@Override
public long getNumRows() {
// Get the number of rows from BQ if not known at this time.
if (numRows == null) {
numRows = BigQuerySQLEngineUtils.getNumRows(bigQuery, bqDataset, bqTable);
}
return numRows;
}
@Override
public String getBigQueryProject() {
return bqDataset.getProject();
}
@Override
public String getBigQueryDataset() {
return bqDataset.getDataset();
}
@Override
public String getBigQueryTable() {
return bqTable;
}
@Override
@Nullable
public String getGCSPath() {
return null;
}
@Override
public String getJobId() {
return jobId;
}
protected void updateTableSchema(TableId tableId, Schema schema) {
// Get BigQuery schema for this table
com.google.cloud.bigquery.Schema bqSchema = BigQuerySinkUtils.convertCdapSchemaToBigQuerySchema(schema);
// Get table and update definition to match the new schema
Table table = bigQuery.getTable(tableId);
TableDefinition updatedDefinition = table.getDefinition().toBuilder().setSchema(bqSchema).build();
Table updatedTable = table.toBuilder().setDefinition(updatedDefinition).build();
// Update table.
bigQuery.update(updatedTable);
}
}