com.bayudwiyansatria.environment.apache.spark.SparkProperties Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of apache-spark Show documentation
Show all versions of apache-spark Show documentation
Bayu Dwiyan Satria For Apache Spark Environment.
The newest version!
/*
* MIT License
*
* Copyright (c) 2019 Bayu Dwiyan Satria
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
* copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in all
* copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NON INFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
package com.bayudwiyansatria.environment.apache.spark;
import com.bayudwiyansatria.Core;
public class SparkProperties extends Core {
/* See https://spark.apache.org/docs/latest/configuration.html for more information */
/* =========================================== Application Properties =========================================== */
// spark.app.name = (none);
private String SPARK_APP_NAME = null;
// spark.driver.cores = 1;
private String SPARK_DRIVER_CORES = null;
// spark.driver.maxResultSize = 1g;
private String SPARK_DRIVER_MAX_RESULT_SIZE = null;
// spark.driver.memory = 1g;
private String SPARK_DRIVER_MEMORY = null;
// spark.driver.memoryOverhead = (none) ex: 0.10;
private String SPARK_DRIVER_MEMORY_OVERHEAD = null;
// spark.driver.allowMultipleContexts = false;
private String SPARK_DRIVER_ALLOW_MULTIPLE_CONTEXT = null;
// spark.executor.memory = 1g;
private String SPARK_EXECUTOR_MEMORY = null;
// spark.executor.pyspark.memory =
private String SPARK_EXECUTOR_PY_SPARK_MEMORY = null;
// spark.executor.memoryOverhead = (none) ex: 0.10;
private String SPARK_EXECUTOR_MEMORY_OVERHEAD = null;
// spark.extraListeners = (none);
private String SPARK_EXTRA_LISTENER = null;
// spark.local.dir = /tmp
private String SPARK_LOCAL_DIR = null;
// spark.logConf = false;
private String SPARK_LOG_CONF = null;
// spark.master = (none);
private String SPARK_MASTER = null;
// spark.submit.deployMode = (none);
private String SPARK_SUBMIT_DEPLOY_MODE = null;
// spark.log.callerContext = (none);
private String SPARK_LOG_CALLER_CONTEXT = null;
// spark.driver.supervise = false;
private String SPARK_DRIVER_SUPERVISE = null;
/* =========================================== Application Properties =========================================== */
/* =========================================== Runtime Environment =========================================== */
// spark.driver.extraClassPath = (none);
private String SPARK_DRIVER_EXTRA_CLASS_PATH = null;
// spark.driver.extraJavaOptions = -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseG1GC -XX:+UseCompressedOops;
private String SPARK_DRIVER_EXTRA_JAVA_OPTION = null;
// spark.driver.extraLibraryPath = (none);
private String SPARK_DRIVER_EXTRA_LIBRARY_PATH = null;
// spark.driver.userClassPathFirst = false;
private String SPARK_DRIVER_USER_CLASS_PATH_FIRST = null;
// spark.executor.extraClassPath = (none);
private String SPARK_EXECUTOR_EXTRA_CLASS_PATH = null;
// spark.executor.extraJavaOptions = (none);
private String SPARK_EXECUTOR_EXTRA_JAVA_OPTION = null;
// spark.executor.extraLibraryPath = (none);
private String SPARK_EXECUTOR_EXTRA_LIBRARY_PATH = null;
// spark.executor.logs.rolling.maxRetainedFiles = (none);
private String SPARK_EXECUTOR_LOGS_ROLLING_MAX_RETAINED_FILES = null;
// spark.executor.logs.rolling.enableCompression = false;
private String SPARK_EXECUTOR_LOGS_ROLLING_ENABLE_COMPRESSION = null;
// spark.executor.logs.rolling.maxSize = (none);
private String SPARK_EXECUTOR_LOGS_ROLLING_MAXSIZE = null;
// spark.executor.logs.rolling.strategy = (none);
private String SPARK_EXECUTOR_LOGS_ROLLING_STRATEGY = null;
// spark.executor.logs.rolling.time.interval = daily;
private String SPARK_EXECUTOR_LOGS_ROLLING_TIME_INTERVAL = null;
// spark.executor.userClassPathFirst = false;
private String SPARK_EXECUTOR_USER_CLASS_PATH_FIRST = null;
// spark.executorEnv.[EnvironmentVariableName] = (none);
private String SPARK_EXECUTOR_ENV = null;
// spark.redaction.regex = (?i)secret|password;
private String SPARK_REDACTION_REGEX = null;
// spark.python.profile = false;
private String SPARK_PYtHON_PROFILE = null;
// spark.python.profile.dump = (none);
private String SPARK_PYTHON_PROFILE_DUMP = null;
// spark.python.worker.memory = 512m;
private String SPARK_PYTHON_WORKER_MEMORY = null;
// spark.python.worker.reuse = true;
private String SPARK_PYTHON_WORKER_REUSE = null;
// spark.files =
private String SPARK_FILES = null;
// spark.submit.pyFiles =
private String SPARK_SUBMIT_PY_FILES = null;
// spark.jars =
private String SPARK_JARS = null;
// spark.jars.packages =
private String SPARK_JARS_PACKAGES = null;
// spark.jars.excludes =
private String SPARK_JARS_EXCLUDES = null;
// spark.jars.ivy =
private String SPARK_JARS_IVY = null;
// spark.jars.ivySettings =
private String SPARK_JARS_IVY_SETTINGS = null;
// spark.jars.repositories =
private String SPARK_JARS_REPOSITORIES = null;
// spark.pyspark.driver.python =
private String SPARK_PY_SPARK_DRIVER_PYTHON = null;
// spark.pyspark.python =
private String SPARK_PY_SPARK_PYTHON = null;
/* =========================================== Runtime Environment =========================================== */
/* =========================================== Shuffle Behavior =========================================== */
// spark.reducer.maxSizeInFlight = 48m;
private String SPARK_REDUCER_MAX_SIZE_IN_FLIGHT = null;
// spark.reducer.maxReqsInFlight = Int.MaxValue;
private String SPARK_REDUCER_MAX_REQ_IN_FLIGHT = null;
// spark.reducer.maxBlocksInFlightPerAddress = Int.MaxValue;
private String SPARK_REDUCER_MAX_BLOCKS_IN_FLIGHT_PER_ADDRESS = null;
// spark.maxRemoteBlockSizeFetchToMem = Int.MaxValue - 512;
private String SPARK_MAX_REMOTE_BLOCK_SIZE_FETCH_TO_MEMORY = null;
// spark.shuffle.compress = true;
private String SPARK_SHUFFLE_COMPRESS = null;
// spark.shuffle.file.buffer = 32k;
private String SPARK_SHUFFLE_FILE_BUFFER = null;
// spark.shuffle.io.maxRetries = 3;
private String SPARK_SHUFFLE_IO_MAX_RETRIES = null;
// spark.shuffle.io.numConnectionsPerPeer = 1;
private String SPARK_SHUFFLE_IO_NUM_CONNECTIONS_PER_PEER = null;
// spark.shuffle.io.preferDirectBufs = true;
private String SPARK_SHUFFLE_IO_PREFER_DIRECT_BUFFERS = null;
// spark.shuffle.io.retryWait = 5s;
private String SPARK_SHUFFLE_IO_RETRY_WAIT = null;
// spark.shuffle.service.enabled = false;
private String SPARK_SHUFFLE_SERVICE_ENABLED = null;
// spark.shuffle.service.port = 7337;
private String SPARK_SHUFFLE_SERVICE_PORT = null;
// spark.shuffle.service.index.cache.size = 100m;
private String SPARK_SHUFFLE_SERVICE_INDEX_CACHE_SIZE = null;
// spark.shuffle.maxChunksBeingTransferred = Long.MAX_VALUE;
private String SPARK_SHUFFLE_MAX_CHUNKS_BEING_TRANSFERRED = null;
// spark.shuffle.sort.bypassMergeThreshold = 200;
private String SPARK_SHUFFLE_SORT_BYPASS_MERGE_THRESHOLD = null;
// spark.shuffle.spill.compress = true;
private String SPARK_SHUFFLE_SPILL_COMPRESS = null;
// spark.shuffle.accurateBlockThreshold = 100 * 1024 * 1024
private String SPARK_SHUFFLE_ACCURATE_BLOCK_THRESHOLD = null;
// spark.shuffle.registration.timeout = 5000;
private String SPARK_SHUFFLE_REGISTRATION_TIMEOUT = null;
// spark.shuffle.registration.maxAttempts = 3;
private String SPARK_SHUFFLE_REGISTRATION_MAXATTEMPTS = null;
/* =========================================== Shuffle Behavior =========================================== */
/* =========================================== Spark UI =========================================== */
// spark.eventLog.logBlockUpdates.enabled = false;
private String SPARK_EVENT_LOG_LOGBLOCKUPDATES_ENABLED = null;
// spark.eventLog.longForm.enabled = false;
private String SPARK_EVENT_LOG_LONGFORM_ENABLED = null;
//spark.eventLog.compress = false;
private String SPARK_EVENT_LOG_COMPRESS = null;
// spark.eventLog.dir = file://tmp/spark-events;
private String SPARK_EVENT_LOG_DIR = null;
// spark.eventLog.enabled = false;
private String SPARK_EVENT_LOG_ENABLED = null;
// spark.eventLog.overwrite = false;
private String SPARK_EVENT_LOG_OVERWRITE = null;
// spark.eventLog.buffer.kb = 100k;
private String SPARK_EVENT_LOG_BUFFER_KB = null;
// spark.ui.dagGraph.retainedRootRDDs = Int.Int.MaxValue;
private String SPARK_UI_DAGGRAPH_RETAINEDROOTRDDS = null;
// spark.ui.enabled = true;
private String SPARK_UI = null;
// spark.ui.killEnabled = true;
private String SPARK_UI_KILL_ENABLED = null;
// spark.ui.liveUpdate.period = 100ms;
private String SPARK_UI_LIVEUPDATEPERIOD = null;
// spark.ui.liveUpdate.minFlushPeriod = 1s;
private String SPARK_UI_LIVEUPDATE_MINFLUSHPERIOD = null;
// spark.ui.port = 4040;
private String SPARK_UI_PORT = null;
// spark.ui.retainedJobs = 1000;
private String SPARK_UI_RETAINEDJOB = null;
// spark.ui.retainedStages = 1000;
private String SPARK_UI_RETAINEDSTAGES = null;
// spark.ui.retainedTasks = 100000;
private String SPARK_UI_RETAINEDTASKS = null;
// spark.ui.reverseProxy = false;
private String SPARK_UI_REVERSEPROXY = null;
// spark.ui.reverseProxyUrl =
private String SPARK_UI_REVERSEPROXYURL = null;
// spark.ui.showConsoleProgress = true;
private String SPARK_UI_SHOWCONSOLEPROGRESS = null;
// spark.worker.ui.retainedExecutors = 1000;
private String SPARK_WORKER_UI_RETAINEDEXECUTORS = null;
// spark.worker.ui.retainedDrivers = 1000;
private String SPARK_WORKER_UI_RETAINEDDRIVER = null;
// spark.sql.ui.retainedExecutions = 1000;
private String SPARK_SQL_UI_RETAINEDEXECUTIONS = null;
// spark.streaming.ui.retainedBatches = 1000;
private String SPARK_STREAMING_UI_RETAINEDBATCHES = null;
// spark.ui.retainedDeadExecutors = 100;
private String SPARK_UI_RETAINEDDEADEXECUTORS = null;
// spark.ui.filters = None;
private String SPARK_UI_FILTERS = null;
// spark.ui.requestHeaderSize = 8k;
private String SPARK_UI_REQUESTHEADERSIZE = null;
/* =========================================== Spark UI =========================================== */
/* =========================================== Compression and Serialization ==================================== */
// spark.broadcast.compress = true;
private String SPARK_BROADCAST_COMPRESS = null;
// spark.io.compression.codec = lz4;
private String SPARK_IO_COMPRESSION_CODEC = null;
// spark.io.compression.lz4.blockSize = 32k;
private String SPARK_IO_COMPRESSION_LZ4_BLOCKSIZE = null;
// spark.io.compression.snappy.blockSize = 32k;
private String SPARK_IO_COMPRESSION_SNAPPY_BLOCKSIZE = null;
// spark.io.compression.zstd.level = 1;
private String SPARK_IO_COMPRESSION_ZSTD_LEVEL = null;
// spark.io.compression.zstd.bufferSize = 32k;
private String SPARK_IO_COMPRESSION_ZSTD_BUFFERSIZE = null;
// spark.kryo.classesToRegister = (none);
private String SPARK_KYRO_CLASSESTOREGISTER = null;
// spark.kryo.referenceTracking = true;
private String SPARK_KYRO_REFERENCETRACKING = null;
// spark.kryo.registrationRequired = false;
private String SPARK_KYRO_REGISTRATIONREQUIRED = null;
// spark.kryo.registrator = (none);
private String SPARK_KYRO_REGISTRATOR = null;
// spark.kryo.unsafe = false;
private String SPARK_KYRO_UNSAFE = null;
// spark.kryoserializer.buffer.max = 64m;
private String SPARK_KYROSERIALIZER_BUFFER_MAX = null;
// spark.kryoserializer.buffer = 64k;
private String SPARK_KYROSERIALIZER_BUFFER = null;
// spark.rdd.compress = false;
private String SPARK_RDD_COMPRESS = null;
//spark.serializer = org.apache.spark.serializer.JavaSerializer
private String SPARK_SERIALIZER = null;
//spark.serializer.objectStreamReset = 100;
private String SPARK_SERIALIZER_OBJECTSTREAMRESET = null;
/* =========================================== Compression and Serialization ==================================== */
/* =========================================== Memory Management ==================================== */
// spark.memory.fraction = 0.6
// spark.memory.storageFraction 0.5
// spark.memory.offHeap.enabled = false;
// spark.memory.offHeap.size = 0;
// spark.memory.useLegacyMode = false;
// spark.shuffle.memoryFraction = 0.2;
// spark.storage.memoryFraction = 0.6;
// spark.storage.unrollFraction = 0.2;
// spark.storage.replication.proactive = false;
// spark.cleaner.periodicGC.interval = 30min;
// spark.cleaner.referenceTracking = true;
// spark.cleaner.referenceTracking.blocking = true;
// spark.cleaner.referenceTracking.blocking.shuffle = false;
// spark.cleaner.referenceTracking.cleanCheckpoints = false;
/* =========================================== Memory Management ==================================== */
/* =========================================== Execution Behavior ==================================== */
// spark.broadcast.blockSize = 4m;
private String SPARK_BROADCAST_BLOCKSIZE = null;
// spark.broadcast.checksum = true;
private String SPARK_BROADCAST_CHECKSUM = null;
// spark.default.parallelism =
private String SPARK_DEFAULT_PARALLELISM = null;
// spark.executor.heartbeatInterval = 10s;
private String SPARK_EXECUTOR_HEARBEATINTERVAL = null;
// spark.files.fetchTimeout = 60s;
private String SPARK_FILES_FETCH_TIMEOUT = null;
// spark.files.useFetchCache = true;
private String SPARK_FILES_USEFETCHCACHE = null;
// spark.files.overwrite = false;
private String SPARK_FILES_OVERWRITE = null;
// spark.files.maxPartitionBytes = 134217728 (128 MB)
private String SPARK_FILES_MAXPARTITIONBYTES = null;
// spark.files.openCostInBytes = 4194304 (4 MB)
private String SPARK_FILES_OPENCOSTINBYTES = null;
// spark.hadoop.cloneConf = false;
private String SPARK_HADOOP_CLONECONF = null;
// spark.hadoop.validateOutputSpecs = true;
private String SPARK_HADOOP_VALIDATEOUTPUTSPECS = null;
// spark.storage.memoryMapThreshold = 2m;
private String SPARK_STORAGE_MEMORYMAPTHRESHOLD = null;
// spark.hadoop.mapreduce.fileoutputcommitter.algorithm.version = 2;
private String SPARK_HADOOP_MAPREDUCE_FILEOUTPUTCOMMUTTER_ALGORITHM_VERSION = null;
/* =========================================== Execution Behavior ==================================== */
/* =========================================== Networking ==================================== */
/* =========================================== Networking ==================================== */
/* =========================================== Scheduling ==================================== */
/* =========================================== Scheduling ==================================== */
/* =========================================== Dynamic Allocation ==================================== */
// spark.dynamicAllocation.enabled = false;
private String SPARK_DYNAMIC_ALLOCATION_ENABLED = null;
// spark.dynamicAllocation.executorIdleTimeout = 60s;
private String SPARK_DYNAMIC_ALLOCATION_EXECUTORIDLETIMEOUT = null;
// spark.dynamicAllocation.cachedExecutorIdleTimeout = infinity;
private String SPARK_DYNAMIC_ALLOCATION_CACHEDEXECUTORIDLETIMEOUT = null;
// spark.dynamicAllocation.initialExecutors = spark.dynamicAllocation.minExecutors;
private String SPARK_DYNAMIC_ALLOCATION_INITIALEXECUTORS = null;
// spark.dynamicAllocation.maxExecutors = infinity;
private String SPARK_DYNAMIC_ALLOCATION_MAXEXECUTORS = null;
// spark.dynamicAllocation.minExecutors = 0;
private String SPARK_DYNAMIC_ALLOCATION_MINEXECUTORS = null;
// spark.dynamicAllocation.executorAllocationRatio = 1;
private String SPARK_DYNAMIC_ALLOCATION_EXECUTORALLOCATIONRATIO = null;
// spark.dynamicAllocation.schedulerBacklogTimeout = 1s;
private String SPARK_DYNAMIC_ALLOCATION_SCHEDULERBACKLOGTIMEOUT = null;
// spark.dynamicAllocation.sustainedSchedulerBacklogTimeout = schedulerBacklogTimeout;
private String SPARK_DYNAMIC_ALLOCATION_SUSTAINEDSCHEDULERBACKLOGTIMEOUT = null;
/* =========================================== Dynamic Allocation ==================================== */
/* =========================================== Security ==================================== */
/* =========================================== Security ==================================== */
/* =========================================== SparkR ==================================== */
/* =========================================== SparkR ==================================== */
/* =========================================== GraphX ==================================== */
/* =========================================== GraphX ==================================== */
/* =========================================== Deploy ==================================== */
/* =========================================== Deploy ==================================== */
/* =========================================== Environment Variables ==================================== */
/* =========================================== Environment Variables ==================================== */
/* =========================================== History Server ======================================== */
// spark.history.provider = org.apache.spark.deploy.history.FsHistoryProvider
private String SPARK_HISTORY_PROVIDER = null;
// spark.history.fs.logDirectory = "file://tmp/spark-events";
private String SPARK_HISTORY_FS_LOGDIRECTORY = null;
// spark.history.fs.update.interval = 10s;
private String SPARK_HISTORY_FS_UPDATE_INTERVAL = null;
// spark.history.retainedApplications = 50;
private String SPARK_HISTORY_RETAINEDAPPLICATIONS = null;
// spark.history.ui.maxApplications = Int.MaxValue;
private String SPARK_HISTORY_UI_MAXAPPLICATION = null;
// spark.history.ui.port = 18080;
private String SPARK_HISTORY_UI_PORT = null;
// spark.history.kerberos.enabled = false;
private String SPARK_HISTORY_KERBEROS_ENABLED = null;
// spark.history.kerberos.principal = (none);
private String SPARK_HISTORY_KERBEROS_PRINCIPAL = null;
// spark.history.kerberos.keytab = (none)
private String SPARK_HISTORY_KERBEROS_KEYTAB = null;
// spark.history.fs.cleaner.enabled = false;
private String SPARK_HISTORY_FS_CLEANER_ENABLED = null;
// spark.history.fs.cleaner.interval = 1d;
private String SPARK_HISTORY_FS_CLEANER_INTERVAL = null;
// spark.history.fs.cleaner.maxAge = 7d;
private String SPARK_HISTORY_FS_CLEANER_MAXAGE = null;
// spark.history.fs.endEventReparseChunkSize = 1m;
private String SPARK_HISTORY_FS_ENDEVENTREPARSECHUNKSIZE = null;
// spark.history.fs.inProgressOptimization.enabled = true;
private String SPARK_HISTORY_FS_INPROGRESSOPTIMAZTION_ENABLED = null;
// spark.history.fs.numReplayThreads = 25% of available cores;
private String SPARK_HISTORY_FS_NUMREPLAYTHREADS = null;
// spark.history.store.maxDiskUsage = 10g;
private String SPARK_HISTORY_STORE_MAXDISKUSAGE = null;
// spark.history.store.path = (none);
private String SPARK_HISTORY_STORE_PATH = null;
/* =========================================== History Server ============================================== */
/* ============================================================================================================== */
/* =========================================== Application Properties =========================================== */
protected void setSparkAppName(String AppName) {
this.SPARK_APP_NAME = "spark.app.name="+AppName;
}
protected String getSparkAppName() {
return SPARK_APP_NAME;
}
protected void setSparkDriverCores(String DriverCores) {
this.SPARK_DRIVER_CORES = "spark.driver.cores="+DriverCores;
}
protected String getSparkDriverCores() {
return SPARK_DRIVER_CORES;
}
public void setSparkDriverMaxResultSize(String ResultSize) {
this.SPARK_DRIVER_MAX_RESULT_SIZE = "spark.driver.maxResultSize="+ResultSize;
}
public String getSparkDriverMaxResultSize() {
return SPARK_DRIVER_MAX_RESULT_SIZE;
}
protected void setSparkDriverMemory(String DriverMemory) {
this.SPARK_DRIVER_MEMORY = "spark.driver.memory="+DriverMemory;
}
protected String getSparkDriverMemory() {
return SPARK_DRIVER_MEMORY;
}
public void setSparkDriverMemoryOverhead(String MemoryOverhead) {
this.SPARK_DRIVER_MEMORY_OVERHEAD = "spark.driver.memoryOverhead="+MemoryOverhead;
}
public String getSparkDriverMemoryOverhead() {
return SPARK_DRIVER_MEMORY_OVERHEAD;
}
public void setSparkDriverAllowMultipleContexts(String AllowMultipleContext) {
this.SPARK_DRIVER_ALLOW_MULTIPLE_CONTEXT = "spark.driver.allowMultipleContexts=" + AllowMultipleContext;
}
public String getSparkDriverAllowMultipleContexts() {
return SPARK_DRIVER_ALLOW_MULTIPLE_CONTEXT;
}
protected void setSparkExecutorMemory(String ExecutorMemory) {
this.SPARK_EXECUTOR_MEMORY = "spark.executor.memory="+ExecutorMemory;
}
protected String getSparkExecutorMemory() {
return SPARK_EXECUTOR_MEMORY;
}
public void setSparkExecutorPySparkMemory(String PySparkMemory) {
this.SPARK_EXECUTOR_PY_SPARK_MEMORY = "spark.executor.pyspark.memory="+PySparkMemory;
}
public String getSparkExecutorPySparkMemory() {
return SPARK_EXECUTOR_PY_SPARK_MEMORY;
}
public void setSparkExecutorMemoryOverhead(String MemoryOverhead) {
this.SPARK_EXECUTOR_MEMORY_OVERHEAD = "spark.executor.memoryOverhead="+MemoryOverhead;
}
public String getSparkExecutorMemoryOverhead() {
return SPARK_EXECUTOR_MEMORY_OVERHEAD;
}
public void setSparkExtraListener(String ExtraListener) {
this.SPARK_EXTRA_LISTENER = "spark.extraListeners="+ExtraListener;
}
public String getSparkExtraListener() {
return SPARK_EXTRA_LISTENER;
}
public void setSparkLocalDirectory(String LocalDirectory) {
this.SPARK_LOCAL_DIR = "spark.local.dir="+LocalDirectory;
}
public String getSparkLocalDirectory() {
return SPARK_LOCAL_DIR;
}
public void setSparkLogConf(String LogConf) {
this.SPARK_LOG_CONF = "spark.logConf="+LogConf;
}
public String getSparkLogCOnf() {
return SPARK_LOG_CONF;
}
protected void setMaster(String SparkMaster) {
this.SPARK_MASTER = "spark.master="+SparkMaster;
}
protected String getMaster() {
return SPARK_MASTER;
}
protected void setSparkSubmitDeployMode(String DeployMode) {
this.SPARK_SUBMIT_DEPLOY_MODE = "spark.submit.deployMode="+DeployMode;
}
protected String getSparkSubmitDeployMode() {
return SPARK_SUBMIT_DEPLOY_MODE;
}
public void setSparkLogCallerContext(String CallerContext) {
this.SPARK_LOG_CALLER_CONTEXT = "spark.log.callerContext="+CallerContext;
}
public String getSparkLogCallerContext() {
return SPARK_LOG_CALLER_CONTEXT;
}
protected void setSparkDriverSupervise(String Supervise) {
this.SPARK_DRIVER_SUPERVISE = "spark.driver.supervise="+Supervise;
}
protected String getSparkDriverSupervise() {
return SPARK_DRIVER_SUPERVISE;
}
/* =========================================== Application Properties =========================================== */
/* =========================================== Runtime Environment =========================================== */
protected void setSparkDriverExtraClassPath(String ExtraClassPath) {
this.SPARK_DRIVER_EXTRA_CLASS_PATH = "spark.driver.extraClassPath=" + ExtraClassPath;
}
protected String getSparkDriverExtraClassPath() {
return SPARK_DRIVER_EXTRA_CLASS_PATH;
}
protected void setSparkDriverExtraJavaOptions(String ExtraJavaOptions) {
this.SPARK_DRIVER_EXTRA_JAVA_OPTION = "spark.driver.extraJavaOptions=" + ExtraJavaOptions;
}
protected String getSparkDriverExtraJavaOptions() {
return SPARK_DRIVER_EXTRA_JAVA_OPTION;
}
protected void setSparkDriverExtraLibraryPath(String ExtraLibraryPath) {
this.SPARK_DRIVER_EXTRA_LIBRARY_PATH = "spark.driver.extraLibraryPath=" + ExtraLibraryPath;
}
protected String getSparkDriverExtraLibraryPath() {
return SPARK_DRIVER_EXTRA_LIBRARY_PATH;
}
protected void setSparkDriverUserClassPathFirst(String UserClassPathFirst) {
this.SPARK_DRIVER_USER_CLASS_PATH_FIRST = "spark.driver.userClassPathFirst=" + UserClassPathFirst;
}
protected String getSparkDriverUserClassPathFirst() {
return SPARK_DRIVER_USER_CLASS_PATH_FIRST;
}
/* =========================================== Runtime Environment =========================================== */
/* =========================================== Shuffle Behavior =========================================== */
public void setSparkReducerMaxSizeInFlight(String ReducerMaxSizeInFlight) {
this.SPARK_REDUCER_MAX_SIZE_IN_FLIGHT = "spark.reducer.maxSizeInFlight=" + ReducerMaxSizeInFlight;
}
public String getSparkReducerMaxSizeInFlight() {
return SPARK_REDUCER_MAX_SIZE_IN_FLIGHT;
}
public void setSparkReducerMaxReqsInFlight(String MaxReqsInFlight) {
this.SPARK_REDUCER_MAX_REQ_IN_FLIGHT = "spark.reducer.maxReqsInFlight=" + MaxReqsInFlight;
}
public String getSparkReducerMaxReqsInFlight() {
return SPARK_REDUCER_MAX_REQ_IN_FLIGHT;
}
public void setSparkReducerMaxBlocksInFlightPerAddress(String MaxBlocksInFlightPerAddress) {
this.SPARK_REDUCER_MAX_BLOCKS_IN_FLIGHT_PER_ADDRESS = "spark.reducer.maxBlocksInFlightPerAddress=" + MaxBlocksInFlightPerAddress;
}
public String getSparkReducerMaxBlocksInFlightPerAddress() {
return SPARK_REDUCER_MAX_BLOCKS_IN_FLIGHT_PER_ADDRESS;
}
public void setSparkMaxRemoteBlockSizeFetchToMem(String MaxRemoteBlockSizeFetchToMem) {
this.SPARK_MAX_REMOTE_BLOCK_SIZE_FETCH_TO_MEMORY = "spark.maxRemoteBlockSizeFetchToMem=" + MaxRemoteBlockSizeFetchToMem;
}
public String getSparkMaxRemoteBlockSizeFetchToMem() {
return SPARK_MAX_REMOTE_BLOCK_SIZE_FETCH_TO_MEMORY;
}
public void setSparkShuffleCompress(String Compress) {
this.SPARK_SHUFFLE_COMPRESS = "spark.shuffle.compress=" + Compress;
}
public String getSparkShuffleCompress() {
return SPARK_SHUFFLE_COMPRESS;
}
public void setSparkShuffleFileBuffer(String FileBuffer) {
this.SPARK_SHUFFLE_FILE_BUFFER = "spark.shuffle.file.buffer=" + FileBuffer;
}
public String getSparkShuffleFileBuffer() {
return SPARK_SHUFFLE_FILE_BUFFER;
}
public void setSparkShuffleIOMaxRetries(String IOMaxRetries) {
this.SPARK_SHUFFLE_IO_MAX_RETRIES = "spark.shuffle.io.maxRetries=" + IOMaxRetries;
}
public String getSparkShuffleIOMaxRetries() {
return SPARK_SHUFFLE_IO_MAX_RETRIES;
}
public void setSparkShuffleIONumConnectionsPerPeer(String IONumConnectionsPerPeer) {
this.SPARK_SHUFFLE_IO_NUM_CONNECTIONS_PER_PEER = "spark.shuffle.io.numConnectionsPerPeer=" + IONumConnectionsPerPeer;
}
public String getSparkShuffleIONumConnectionsPerPeer() {
return SPARK_SHUFFLE_IO_NUM_CONNECTIONS_PER_PEER;
}
public void setSparkShuffleIOPreferDirectBufs(String IOPreferDirectBufs) {
this.SPARK_SHUFFLE_IO_PREFER_DIRECT_BUFFERS = "spark.shuffle.io.preferDirectBufs=" + IOPreferDirectBufs;
}
public String getSparkShuffleIOPreferDirectBufs() {
return SPARK_SHUFFLE_IO_PREFER_DIRECT_BUFFERS;
}
public void setSparkShuffleIORetryWait(String RetryWait) {
this.SPARK_SHUFFLE_IO_RETRY_WAIT = "spark.shuffle.io.retryWait=" + RetryWait;
}
public String getSparkShuffleIORetryWait() {
return SPARK_SHUFFLE_IO_RETRY_WAIT;
}
public void setSparkShuffleServiceEnabled(String ServiceEnabled) {
this.SPARK_SHUFFLE_SERVICE_ENABLED = "spark.shuffle.service.enabled=" + ServiceEnabled;
}
public String getSparkShuffleServiceEnabled() {
return SPARK_SHUFFLE_SERVICE_ENABLED;
}
public void setSparkShuffleServicePort(String ServicePort) {
this.SPARK_SHUFFLE_SERVICE_PORT = "spark.shuffle.service.port=" + ServicePort;
}
public String getSparkShuffleServicePort() {
return SPARK_SHUFFLE_SERVICE_PORT;
}
public void setSparkShuffleServiceIndexCacheSize(String ServiceIndexCacheSize) {
this.SPARK_SHUFFLE_SERVICE_INDEX_CACHE_SIZE = "spark.shuffle.service.index.cache.size=" + ServiceIndexCacheSize;
}
public String getSparkShuffleServiceIndexCacheSize() {
return SPARK_SHUFFLE_SERVICE_INDEX_CACHE_SIZE;
}
public void setSparkShuffleMaxChunksBeingTransferred(String MaxChunksBeingTransferred) {
this.SPARK_SHUFFLE_MAX_CHUNKS_BEING_TRANSFERRED = "spark.shuffle.maxChunksBeingTransferred=" + MaxChunksBeingTransferred;
}
public String getSparkShuffleMaxChunksBeingTransferred() {
return SPARK_SHUFFLE_MAX_CHUNKS_BEING_TRANSFERRED;
}
public void setSparkShuffleSortBypassMergeThreshold(String SortBypassMergeThreshold) {
this.SPARK_SHUFFLE_SORT_BYPASS_MERGE_THRESHOLD = "spark.shuffle.sort.bypassMergeThreshold=" + SortBypassMergeThreshold;
}
public String getSparkShuffleSortBypassMergeThreshold() {
return SPARK_SHUFFLE_SORT_BYPASS_MERGE_THRESHOLD;
}
public void setSparkShuffleSpillCompress(String SpillCompress) {
this.SPARK_SHUFFLE_SPILL_COMPRESS = "spark.shuffle.spill.compress=" + SpillCompress;
}
public String getSparkShuffleSpillCompress() {
return SPARK_SHUFFLE_SPILL_COMPRESS;
}
public void setSparkShuffleAccurateBlockThreshold(String AccurateBlockThreshold) {
this.SPARK_SHUFFLE_ACCURATE_BLOCK_THRESHOLD = "spark.shuffle.accurateBlockThreshold=" + AccurateBlockThreshold;
}
public String getSparkShuffleAccurateBlockThreshold() {
return SPARK_SHUFFLE_ACCURATE_BLOCK_THRESHOLD;
}
public void setSparkShuffleRegistrationTimeout(String RegistrationTimeout) {
this.SPARK_SHUFFLE_REGISTRATION_TIMEOUT = "spark.shuffle.registration.timeout=" + RegistrationTimeout;
}
public String getSparkShuffleRegistrationTimeout() {
return SPARK_SHUFFLE_REGISTRATION_TIMEOUT;
}
public void setSparkShuffleRegistrationMaxAttempts(String RegistrationMaxAttempts) {
this.SPARK_SHUFFLE_REGISTRATION_MAXATTEMPTS = "spark.shuffle.registration.maxAttempts=" + RegistrationMaxAttempts;
}
public String getSparkShuffleRegistrationMaxAttempts() {
return SPARK_SHUFFLE_REGISTRATION_MAXATTEMPTS;
}
/* =========================================== Shuffle Behavior =========================================== */
/* =========================================== Spark UI =========================================== */
public void setSparkEventLogBlockUpdatesEnabled(String BlockUpdatesEnabled) {
this.SPARK_EVENT_LOG_LOGBLOCKUPDATES_ENABLED = "spark.eventLog.logBlockUpdates.enabled=" + BlockUpdatesEnabled;
}
public String getSparkEventLogBlockUpdatesEnabled() {
return SPARK_EVENT_LOG_LOGBLOCKUPDATES_ENABLED;
}
public void setSparkEventLogLongFormEnabled(String LongFormEnabled) {
this.SPARK_EVENT_LOG_LONGFORM_ENABLED = "spark.eventLog.longForm.enabled=" + LongFormEnabled;
}
public String getSparkEventLogLongFormEnabled() {
return SPARK_EVENT_LOG_LONGFORM_ENABLED;
}
public void setSparkEventLogCompress(String LogCompress) {
this.SPARK_EVENT_LOG_COMPRESS = "spark.eventLog.compress=" + LogCompress;
}
public String getSparkEventLogCompress() {
return SPARK_EVENT_LOG_COMPRESS;
}
public void setSparkEventLogDir(String EventLogDir) {
this.SPARK_EVENT_LOG_DIR = "spark.eventLog.dir=" + EventLogDir;
}
public String getSparkEventLogDir() {
return SPARK_EVENT_LOG_DIR;
}
public void setSparkEventLogEnabled(String EventLogEnabled) {
this.SPARK_EVENT_LOG_ENABLED = "spark.eventLog.enabled=" + EventLogEnabled;
}
public String getSparkEventLogEnabled() {
return SPARK_EVENT_LOG_ENABLED;
}
public void setSparkEventLogOverWrite(String EventLogOverWrite) {
this.SPARK_EVENT_LOG_OVERWRITE = "spark.eventLog.overwrite=" + EventLogOverWrite;
}
public String getSparkEventLogOverWrite() {
return SPARK_EVENT_LOG_OVERWRITE;
}
public void setSparkEventLogBufferKB(String BufferKB) {
this.SPARK_EVENT_LOG_BUFFER_KB = "spark.eventLog.buffer.kb=" + BufferKB;
}
public String getSparkEventLogBufferKB() {
return SPARK_EVENT_LOG_BUFFER_KB;
}
/* =========================================== Spark UI =========================================== */
/* =========================================== Compression and Serialization ==================================== */
public void setSparkBroadcastCompress(String SparkBroadcastCompress) {
this.SPARK_BROADCAST_COMPRESS = "spark.broadcast.compress=" + SparkBroadcastCompress;
}
public String getSparkBroadcastCompress() {
return SPARK_BROADCAST_COMPRESS;
}
public void setSparkIOCompressionCodec(String setSparkIOCompressionCodec) {
this.SPARK_IO_COMPRESSION_CODEC = "spark.io.compression.codec=" + setSparkIOCompressionCodec;
}
public String getSparkIOCompressionCodec() {
return SPARK_IO_COMPRESSION_CODEC;
}
public void setSparkIOCompressionLz4Blocksize(String CompressionLz4BlockSize) {
this.SPARK_IO_COMPRESSION_LZ4_BLOCKSIZE = "spark.io.compression.lz4.blockSize=" + CompressionLz4BlockSize;
}
public String getSparkIOCompressionLz4Blocksize() {
return SPARK_IO_COMPRESSION_LZ4_BLOCKSIZE;
}
public void setSparkIOCompressionSnappyBlockSize(String CompressionSnappyBlockSize) {
this.SPARK_IO_COMPRESSION_SNAPPY_BLOCKSIZE = "spark.io.compression.snappy.blockSize=" + CompressionSnappyBlockSize;
}
public String getSparkIOCompressionSnappyBlockSize() {
return SPARK_IO_COMPRESSION_SNAPPY_BLOCKSIZE;
}
public void setSparkIOCompressionZSTDLevel(String CompressionZSTDLevel) {
this.SPARK_IO_COMPRESSION_ZSTD_LEVEL = "spark.io.compression.zstd.level=" + CompressionZSTDLevel;
}
public String getSparkIOCompressionZSTDLevel() {
return SPARK_IO_COMPRESSION_ZSTD_LEVEL;
}
public void setSparkIOCompressionZSTDBufferSize(String CompressionZSTDBufferSize) {
this.SPARK_IO_COMPRESSION_ZSTD_BUFFERSIZE = "spark.io.compression.zstd.bufferSize=" + CompressionZSTDBufferSize;
}
public String getSparkIOCompressionZSTDBufferSize() {
return SPARK_IO_COMPRESSION_ZSTD_BUFFERSIZE;
}
public void setKyroClassToRegister(String KyroClassToRegister) {
this.SPARK_KYRO_CLASSESTOREGISTER = "spark.kryo.classesToRegister=" + KyroClassToRegister;
}
public String getKyroClassToRegister() {
return SPARK_KYRO_CLASSESTOREGISTER;
}
public void setKyroReferenceTracking(String ReferenceTracking) {
this.SPARK_KYRO_REFERENCETRACKING = "spark.kryo.referenceTracking=" + ReferenceTracking;
}
public String getKyroReferenceTracking() {
return SPARK_KYRO_REFERENCETRACKING;
}
public void setSparkKyroRequired(String KyroRegisterRequired) {
this.SPARK_KYRO_REGISTRATIONREQUIRED = "spark.kryo.registrationRequired=" + KyroRegisterRequired;
}
public String getSparkKyroRequired() {
return SPARK_KYRO_REGISTRATIONREQUIRED;
}
public void setSparkKyroRegistrator(String KyroRegistrar) {
this.SPARK_KYRO_REGISTRATOR = "spark.kryo.registrator=" + KyroRegistrar;
}
public String getSparkKyroRegistrator() {
return SPARK_KYRO_REGISTRATOR;
}
public void setSparkKyroUnsafe(String Unsafe) {
this.SPARK_KYRO_UNSAFE = "spark.kryo.unsafe=" + Unsafe;
}
public String getSparkKyroUnsafe() {
return SPARK_KYRO_UNSAFE;
}
public void setSparkKyroSerializerBufferMax(String SerializerBufferMax) {
this.SPARK_KYROSERIALIZER_BUFFER_MAX = "spark.kryoserializer.buffer.max=" + SerializerBufferMax;
}
public String getSparkKyroSerializerBufferMax() {
return SPARK_KYROSERIALIZER_BUFFER_MAX;
}
public void setKyroSerializerBuffer(String SerializerBuffer) {
this.SPARK_KYROSERIALIZER_BUFFER = "spark.kryoserializer.buffer=" + SerializerBuffer;
}
public String getSparkKyroSerializerBuffer() {
return SPARK_KYROSERIALIZER_BUFFER;
}
public void setRDDCompress(String RDDCompress) {
this.SPARK_RDD_COMPRESS = "spark.rdd.compress=" + RDDCompress;
}
public String getRDDCompress() {
return SPARK_RDD_COMPRESS;
}
public void setSparkSerializer(String SparkSerializer) {
this.SPARK_SERIALIZER = "spark.serializer=" + SparkSerializer;
}
public String getSparkSerializer() {
return SPARK_SERIALIZER;
}
public void setSparkSerializerObjectStreamReset(String SerializerObjectStreamReset) {
this.SPARK_SERIALIZER_OBJECTSTREAMRESET = "spark.serializer.objectStreamReset=" + SerializerObjectStreamReset;
}
public String getSparkSerializerObjectStreamReset() {
return SPARK_SERIALIZER_OBJECTSTREAMRESET;
}
/* =========================================== Compression and Serialization ==================================== */
/* =========================================== Memory Management ==================================== */
/* =========================================== Memory Management ==================================== */
/* =========================================== Execution Behavior ==================================== */
public void setSparkDefaultParallelism(String DefaultParallelism) {
this.SPARK_DEFAULT_PARALLELISM = "spark.default.parallelism=" + DefaultParallelism;
}
public String getSparkDefaultParallelism() {
return SPARK_DEFAULT_PARALLELISM;
}
/* =========================================== Execution Behavior ==================================== */
/* =========================================== Networking ==================================== */
/* =========================================== Networking ==================================== */
/* =========================================== Scheduling ==================================== */
/* =========================================== Scheduling ==================================== */
/* =========================================== Dynamic Allocation ==================================== */
public void setSparkDynamicAllocationEnabled(String DynamicAllocationEnabled) {
this.SPARK_DYNAMIC_ALLOCATION_ENABLED = "spark.dynamicAllocation.enabled=" + DynamicAllocationEnabled;
}
public String getSparkDynamicAllocationEnabled() {
return SPARK_DYNAMIC_ALLOCATION_ENABLED;
}
public void setSparkDynamicAllocationExecutorIdleTimeout(String ExecutorIdleTimeout) {
this.SPARK_DYNAMIC_ALLOCATION_EXECUTORIDLETIMEOUT = "spark.dynamicAllocation.executorIdleTimeout=" + ExecutorIdleTimeout;
}
public String getSparkDynamicAllocationExecutorIdleTimeout() {
return SPARK_DYNAMIC_ALLOCATION_EXECUTORIDLETIMEOUT;
}
public void setSparkDynamicAllocationCachedExecutorIdleTimeout(String CachedExecutorIdleTimeout) {
this.SPARK_DYNAMIC_ALLOCATION_CACHEDEXECUTORIDLETIMEOUT = "spark.dynamicAllocation.cachedExecutorIdleTimeout=" + CachedExecutorIdleTimeout;
}
public String getSparkDynamicAllocationCachedExecutorIdleTimeout() {
return SPARK_DYNAMIC_ALLOCATION_CACHEDEXECUTORIDLETIMEOUT;
}
public void setSparkDynamicAllocationInitialExecutors(String InitialExecutors) {
this.SPARK_DYNAMIC_ALLOCATION_INITIALEXECUTORS = "spark.dynamicAllocation.initialExecutors=" + InitialExecutors;
}
public String getSparkDynamicAllocationInitialExecutors() {
return SPARK_DYNAMIC_ALLOCATION_INITIALEXECUTORS;
}
public void setSparkDynamicAllocationMaxExecutors(String MaxExecutors) {
this.SPARK_DYNAMIC_ALLOCATION_MAXEXECUTORS = "spark.dynamicAllocation.maxExecutors=" + MaxExecutors;
}
public String getSparkDynamicAllocationMaxExecutors() {
return SPARK_DYNAMIC_ALLOCATION_MAXEXECUTORS;
}
public void setSparkDynamicAllocationMinExecutors(String MinExecutors) {
this.SPARK_DYNAMIC_ALLOCATION_MINEXECUTORS = "spark.dynamicAllocation.minExecutors=" + MinExecutors;
}
public String getSparkDynamicAllocationMinExecutors() {
return SPARK_DYNAMIC_ALLOCATION_MINEXECUTORS;
}
public void setDynamicAllocationExecutorAllocationRatio(String ExecutorAllocationRation) {
this.SPARK_DYNAMIC_ALLOCATION_EXECUTORALLOCATIONRATIO = "spark.dynamicAllocation.executorAllocationRatio=" + ExecutorAllocationRation;
}
public String getDynamicAllocationExecutorAllocationRatio() {
return SPARK_DYNAMIC_ALLOCATION_EXECUTORALLOCATIONRATIO;
}
public void setDynamicAllocationSchedulerBacklogTimeout(String SchedulerBacklogTimeout) {
this.SPARK_DYNAMIC_ALLOCATION_SCHEDULERBACKLOGTIMEOUT = "spark.dynamicAllocation.schedulerBacklogTimeout=" + SchedulerBacklogTimeout;
}
public String getDynamicAllocationSchedulerBacklogTimeout() {
return SPARK_DYNAMIC_ALLOCATION_SCHEDULERBACKLOGTIMEOUT;
}
public void setDynamicAllocationSustainedSchedulerBacklogTimeout(String SustainedSchedulerBacklogTimeout) {
this.SPARK_DYNAMIC_ALLOCATION_SUSTAINEDSCHEDULERBACKLOGTIMEOUT = "spark.dynamicAllocation.sustainedSchedulerBacklogTimeout=" + SustainedSchedulerBacklogTimeout;
}
public String getDynamicAllocationSustainedSchedulerBacklogTimeout() {
return SPARK_DYNAMIC_ALLOCATION_SUSTAINEDSCHEDULERBACKLOGTIMEOUT;
}
/* =========================================== Dynamic Allocation ==================================== */
/* =========================================== Security ==================================== */
/* =========================================== Security ==================================== */
/* =========================================== SparkR ==================================== */
/* =========================================== SparkR ==================================== */
/* =========================================== GraphX ==================================== */
/* =========================================== GraphX ==================================== */
/* =========================================== Deploy ==================================== */
/* =========================================== Deploy ==================================== */
/* =========================================== Environment Variables ==================================== */
/* =========================================== Environment Variables ==================================== */
/* =========================================== History Server ======================================== */
public void setSparkHistoryProvider(String HistoryProvider) {
this.SPARK_HISTORY_PROVIDER = "spark.history.provider=" + HistoryProvider;
}
public String getSparkHistoryProvider() {
return SPARK_HISTORY_PROVIDER;
}
public void setSparkHistoryFSLogDirectory(String LogDirectory) {
this.SPARK_HISTORY_FS_LOGDIRECTORY = "spark.history.fs.logDirectory=" + LogDirectory;
}
public String getSparkHistoryFSLogDirectory() {
return SPARK_HISTORY_FS_LOGDIRECTORY;
}
public void setSparkHistoryFSUpdateInterval(String UpdateInterval) {
this.SPARK_HISTORY_FS_UPDATE_INTERVAL = "spark.history.fs.update.interval=" + UpdateInterval;
}
public String getSparkHistoryFSUpdateInterval() {
return SPARK_HISTORY_FS_UPDATE_INTERVAL;
}
public void setSparkHistoryRetainedApplications(String RetainedApplications) {
this.SPARK_HISTORY_FS_UPDATE_INTERVAL = "spark.history.retainedApplications=" + RetainedApplications;
}
public String getSparkHistoryRetainedApplications() {
return SPARK_HISTORY_RETAINEDAPPLICATIONS;
}
public void setSparkHistoryUiMaxApplications(String MaxApplication) {
this.SPARK_HISTORY_UI_MAXAPPLICATION = "spark.history.ui.maxApplications=" + MaxApplication;
}
public String getSparkHistoryUiMaxApplications() {
return SPARK_HISTORY_UI_MAXAPPLICATION;
}
public void setSparkHistoryUIPort(String UIPort) {
this.SPARK_HISTORY_UI_PORT = "spark.history.ui.port=" + UIPort;
}
public String getSparkHistoryUIPort() {
return SPARK_HISTORY_UI_PORT;
}
public void setSparkHistoryFsCleanerEnabled(String CleanerEnabled) {
this.SPARK_HISTORY_FS_CLEANER_ENABLED = " spark.history.fs.cleaner.enabled=" + CleanerEnabled;
}
public String getSparkHistoryFsCleanerEnabled() {
return SPARK_HISTORY_FS_CLEANER_ENABLED;
}
public void setSparkHistoryFsCleanerInterval(String CleanerInterval) {
this.SPARK_HISTORY_FS_CLEANER_INTERVAL = "spark.history.fs.cleaner.interval=" + CleanerInterval;
}
public String getSparkHistoryFsCleanerInterval() {
return SPARK_HISTORY_FS_CLEANER_INTERVAL;
}
public void setSparkHistoryFsCleanerMaxAge(String MaxAge) {
this.SPARK_HISTORY_FS_CLEANER_MAXAGE = "spark.history.fs.cleaner.maxAge=" + MaxAge;
}
public String getSparkHistoryFsCleanerMaxAge() {
return SPARK_HISTORY_FS_CLEANER_MAXAGE;
}
public void setSparkHistoryFsEndEventReparseChunkSize(String EndEnvenReparseChunkSize) {
this.SPARK_HISTORY_FS_ENDEVENTREPARSECHUNKSIZE = " spark.history.fs.endEventReparseChunkSize=" + EndEnvenReparseChunkSize;
}
public String getSparkHistoryFsEndEventReparseChunkSize() {
return SPARK_HISTORY_FS_ENDEVENTREPARSECHUNKSIZE;
}
public void setSparkHistoryFsInProgressOptimizationEnabled(String OptimizationEnabled) {
this.SPARK_HISTORY_FS_INPROGRESSOPTIMAZTION_ENABLED = "spark.history.fs.inProgressOptimization.enabled=" + OptimizationEnabled;
}
public String getSparkHistoryFsInProgressOptimizationEnabled() {
return SPARK_HISTORY_FS_INPROGRESSOPTIMAZTION_ENABLED;
}
public void setSparkHistoryFsNumReplayThreads(String NumberReplayThreads) {
this.SPARK_HISTORY_FS_NUMREPLAYTHREADS = "spark.history.fs.numReplayThreads=" + NumberReplayThreads;
}
public String getSparkHistoryFsNumReplayThreads() {
return SPARK_HISTORY_FS_NUMREPLAYTHREADS;
}
public void setSparkHistoryStoreMaxDiskUsage(String MaxDiskUsage) {
this.SPARK_HISTORY_STORE_MAXDISKUSAGE = "spark.history.store.maxDiskUsage=" + MaxDiskUsage;
}
public String getSparkHistoryStoreMaxDiskUsage() {
return SPARK_HISTORY_STORE_MAXDISKUSAGE;
}
public void setSparkHistoryStorePath(String StorePath) {
this.SPARK_HISTORY_STORE_PATH = "spark.history.store.path=" + StorePath;
}
public String getSparkHistoryStorePath() {
return SPARK_HISTORY_STORE_PATH;
}
/* =========================================== History Server ============================================== */
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy