All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.tigergraph.spark.write.TigerGraphBatchWrite Maven / Gradle / Ivy

There is a newer version: 0.2.1
Show newest version
/**
 * Copyright (c) 2023 TigerGraph Inc.
 *
 * 

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file * except in compliance with the License. You may obtain a copy of the License at * *

http://www.apache.org/licenses/LICENSE-2.0 * *

Unless required by applicable law or agreed to in writing, software distributed under the * License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either * express or implied. See the License for the specific language governing permissions and * limitations under the License. */ package com.tigergraph.spark.write; import org.apache.spark.sql.connector.write.BatchWrite; import org.apache.spark.sql.connector.write.DataWriter; import org.apache.spark.sql.connector.write.PhysicalWriteInfo; import org.apache.spark.sql.connector.write.WriterCommitMessage; import org.apache.spark.sql.types.StructType; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import com.tigergraph.spark.TigerGraphConnection; import com.tigergraph.spark.client.common.RestppResponse; import com.tigergraph.spark.util.Utils; /** * Define how to write the data to TG for batch processing. * *

The writing procedure is: * *

    *
  1. Create a writer factory by {@link #createBatchWriterFactory(PhysicalWriteInfo)}, serialize * and send it to all the partitions of the input data(RDD). *
  2. For each partition, create the data writer, and write the data of the partition with this * writer. If all the data are written successfully, call {@link DataWriter#commit()}. If * exception happens during the writing, call {@link DataWriter#abort()}. *
  3. If all writers are successfully committed, call {@link #commit(WriterCommitMessage[])}. If * some writers are aborted, or the job failed with an unknown reason, call {@link * #abort(WriterCommitMessage[])}. *
* *

*/ public class TigerGraphBatchWrite extends TigerGraphWriteBase implements BatchWrite { private static final Logger logger = LoggerFactory.getLogger(TigerGraphBatchWrite.class); TigerGraphBatchWrite(StructType schema, TigerGraphConnection conn) { super(schema, conn); } @Override public TigerGraphBatchWriterFactory createBatchWriterFactory(PhysicalWriteInfo info) { return new TigerGraphBatchWriterFactory(schema, conn); } @Override public void commit(WriterCommitMessage[] messages) { logger.info( "Finished batch loading job {}", conn.getLoadingJobId() == null ? "" : conn.getLoadingJobId()); logger.info("Total processed rows: {}", getTotalProcessedRows(messages)); logger.info("Processed rows of each task:\n{}", getTaskSummury(messages)); RestppResponse resp = getLoadingStatistics(); if (resp != null) { Utils.removeUserData(resp.results); logger.info("Overall loading statistics: {}", resp.results.toPrettyString()); } } @Override public void abort(WriterCommitMessage[] messages) { logger.error( "Aborted batch loading job {}", conn.getLoadingJobId() == null ? "" : conn.getLoadingJobId()); logger.info("Total processed rows: {}", getTotalProcessedRows(messages)); logger.info("Processed rows of each task:\n{}", getTaskSummury(messages)); RestppResponse resp = getLoadingStatistics(); if (resp != null) { Utils.removeUserData(resp.results); logger.info("Overall loading statistics: {}", resp.results.toPrettyString()); } } }





© 2015 - 2024 Weber Informatics LLC | Privacy Policy