Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hudi.common.engine;
import org.apache.hudi.common.data.HoodieAccumulator;
import org.apache.hudi.common.data.HoodieData;
import org.apache.hudi.common.data.HoodieData.HoodieDataCacheKey;
import org.apache.hudi.common.function.SerializableBiFunction;
import org.apache.hudi.common.function.SerializableConsumer;
import org.apache.hudi.common.function.SerializableFunction;
import org.apache.hudi.common.function.SerializablePairFlatMapFunction;
import org.apache.hudi.common.function.SerializablePairFunction;
import org.apache.hudi.common.util.Functions;
import org.apache.hudi.common.util.Option;
import org.apache.hudi.common.util.collection.ImmutablePair;
import org.apache.hudi.common.util.collection.Pair;
import org.apache.hudi.storage.StorageConfiguration;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.stream.Stream;
/**
* Base class contains the context information needed by the engine at runtime. It will be extended by different
* engine implementation if needed.
*/
public abstract class HoodieEngineContext {
/**
* A wrapped hadoop configuration which can be serialized.
*/
private StorageConfiguration> storageConf;
protected TaskContextSupplier taskContextSupplier;
public HoodieEngineContext(StorageConfiguration> storageConf, TaskContextSupplier taskContextSupplier) {
this.storageConf = storageConf;
this.taskContextSupplier = taskContextSupplier;
}
public StorageConfiguration> getStorageConf() {
return storageConf;
}
public TaskContextSupplier getTaskContextSupplier() {
return taskContextSupplier;
}
public abstract HoodieAccumulator newAccumulator();
public abstract HoodieData emptyHoodieData();
public HoodieData parallelize(List data) {
if (data.isEmpty()) {
return emptyHoodieData();
} else {
return parallelize(data, data.size());
}
}
public abstract HoodieData parallelize(List data, int parallelism);
public abstract List map(List data, SerializableFunction func, int parallelism);
public abstract List mapToPairAndReduceByKey(
List data, SerializablePairFunction mapToPairFunc, SerializableBiFunction reduceFunc, int parallelism);
public abstract Stream> mapPartitionsToPairAndReduceByKey(
Stream data, SerializablePairFlatMapFunction, K, V> flatMapToPairFunc,
SerializableBiFunction reduceFunc, int parallelism);
public abstract List reduceByKey(
List> data, SerializableBiFunction reduceFunc, int parallelism);
public abstract List flatMap(List data, SerializableFunction> func, int parallelism);
public abstract void foreach(List data, SerializableConsumer consumer, int parallelism);
public abstract Map mapToPair(List data, SerializablePairFunction func, Integer parallelism);
public abstract void setProperty(EngineProperty key, String value);
public abstract Option getProperty(EngineProperty key);
public abstract void setJobStatus(String activeModule, String activityDescription);
public abstract void putCachedDataIds(HoodieDataCacheKey cacheKey, int... ids);
public abstract List getCachedDataIds(HoodieDataCacheKey cacheKey);
public abstract List removeCachedDataIds(HoodieDataCacheKey cacheKey);
public abstract void cancelJob(String jobId);
public abstract void cancelAllJobs();
/**
* Aggregate the elements of each partition, and then the results for all the partitions, using given combine functions and a neutral "zero value".
*
* @param data
* @param zeroValue the initial value for the accumulated result of each partition for the seqOp operator
* @param seqOp function to aggregate the elements of each partition
* @param combOp function to combine results from different partitions
* @param input object type
* @param output object type
* @return the result of the aggregation
*/
public abstract O aggregate(HoodieData data, O zeroValue, Functions.Function2 seqOp, Functions.Function2 combOp);
}