Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
com.pulumi.gcp.dataproc.kotlin.inputs.WorkflowTemplateJobPysparkJobArgs.kt Maven / Gradle / Ivy
Go to download
Build cloud applications and infrastructure by combining the safety and reliability of infrastructure as code with the power of the Kotlin programming language.
@file:Suppress("NAME_SHADOWING", "DEPRECATION")
package com.pulumi.gcp.dataproc.kotlin.inputs
import com.pulumi.core.Output
import com.pulumi.core.Output.of
import com.pulumi.gcp.dataproc.inputs.WorkflowTemplateJobPysparkJobArgs.builder
import com.pulumi.kotlin.ConvertibleToJava
import com.pulumi.kotlin.PulumiNullFieldException
import com.pulumi.kotlin.PulumiTagMarker
import com.pulumi.kotlin.applySuspend
import kotlin.Pair
import kotlin.String
import kotlin.Suppress
import kotlin.Unit
import kotlin.collections.List
import kotlin.collections.Map
import kotlin.jvm.JvmName
/**
*
* @property archiveUris HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
* @property args The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
* @property fileUris HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
* @property jarFileUris HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
* @property loggingConfig The runtime log config for job execution.
* @property mainPythonFileUri Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
* @property properties A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
* @property pythonFileUris HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
*/
public data class WorkflowTemplateJobPysparkJobArgs(
public val archiveUris: Output>? = null,
public val args: Output>? = null,
public val fileUris: Output>? = null,
public val jarFileUris: Output>? = null,
public val loggingConfig: Output? = null,
public val mainPythonFileUri: Output,
public val properties: Output>? = null,
public val pythonFileUris: Output>? = null,
) : ConvertibleToJava {
override fun toJava(): com.pulumi.gcp.dataproc.inputs.WorkflowTemplateJobPysparkJobArgs =
com.pulumi.gcp.dataproc.inputs.WorkflowTemplateJobPysparkJobArgs.builder()
.archiveUris(archiveUris?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
.args(args?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
.fileUris(fileUris?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
.jarFileUris(jarFileUris?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
.loggingConfig(loggingConfig?.applyValue({ args0 -> args0.let({ args0 -> args0.toJava() }) }))
.mainPythonFileUri(mainPythonFileUri.applyValue({ args0 -> args0 }))
.properties(
properties?.applyValue({ args0 ->
args0.map({ args0 ->
args0.key.to(args0.value)
}).toMap()
}),
)
.pythonFileUris(pythonFileUris?.applyValue({ args0 -> args0.map({ args0 -> args0 }) })).build()
}
/**
* Builder for [WorkflowTemplateJobPysparkJobArgs].
*/
@PulumiTagMarker
public class WorkflowTemplateJobPysparkJobArgsBuilder internal constructor() {
private var archiveUris: Output>? = null
private var args: Output>? = null
private var fileUris: Output>? = null
private var jarFileUris: Output>? = null
private var loggingConfig: Output? = null
private var mainPythonFileUri: Output? = null
private var properties: Output>? = null
private var pythonFileUris: Output>? = null
/**
* @param value HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
*/
@JvmName("ehgycshymtglwlbt")
public suspend fun archiveUris(`value`: Output>) {
this.archiveUris = value
}
@JvmName("imbbifpojmsoetda")
public suspend fun archiveUris(vararg values: Output) {
this.archiveUris = Output.all(values.asList())
}
/**
* @param values HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
*/
@JvmName("hdlbrjbbpfvywloj")
public suspend fun archiveUris(values: List>) {
this.archiveUris = Output.all(values)
}
/**
* @param value The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
*/
@JvmName("wmkjesdxhbdvcrob")
public suspend fun args(`value`: Output>) {
this.args = value
}
@JvmName("owlvwidubmtpcvkp")
public suspend fun args(vararg values: Output) {
this.args = Output.all(values.asList())
}
/**
* @param values The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
*/
@JvmName("qltajnsvvqbkircv")
public suspend fun args(values: List>) {
this.args = Output.all(values)
}
/**
* @param value HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
*/
@JvmName("traufpajngqybpjr")
public suspend fun fileUris(`value`: Output>) {
this.fileUris = value
}
@JvmName("kfiypxhqxcqacxow")
public suspend fun fileUris(vararg values: Output) {
this.fileUris = Output.all(values.asList())
}
/**
* @param values HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
*/
@JvmName("jnjilghxqtdaquws")
public suspend fun fileUris(values: List>) {
this.fileUris = Output.all(values)
}
/**
* @param value HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
*/
@JvmName("uijoxmcjwqchgihq")
public suspend fun jarFileUris(`value`: Output>) {
this.jarFileUris = value
}
@JvmName("atmiwubvxvvxkugt")
public suspend fun jarFileUris(vararg values: Output) {
this.jarFileUris = Output.all(values.asList())
}
/**
* @param values HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
*/
@JvmName("jfgxqbviioxgpvgp")
public suspend fun jarFileUris(values: List>) {
this.jarFileUris = Output.all(values)
}
/**
* @param value The runtime log config for job execution.
*/
@JvmName("twubdwirryqutxfo")
public suspend fun loggingConfig(`value`: Output) {
this.loggingConfig = value
}
/**
* @param value Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
*/
@JvmName("qljbwhrqhfaxarwl")
public suspend fun mainPythonFileUri(`value`: Output) {
this.mainPythonFileUri = value
}
/**
* @param value A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
*/
@JvmName("ylmqcqgafpskhxgi")
public suspend fun properties(`value`: Output>) {
this.properties = value
}
/**
* @param value HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
*/
@JvmName("igyioxyctymxnohx")
public suspend fun pythonFileUris(`value`: Output>) {
this.pythonFileUris = value
}
@JvmName("clbnyxaditytptpl")
public suspend fun pythonFileUris(vararg values: Output) {
this.pythonFileUris = Output.all(values.asList())
}
/**
* @param values HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
*/
@JvmName("bpjtamdaqngitnkm")
public suspend fun pythonFileUris(values: List>) {
this.pythonFileUris = Output.all(values)
}
/**
* @param value HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
*/
@JvmName("nesvfwwuwalejtah")
public suspend fun archiveUris(`value`: List?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.archiveUris = mapped
}
/**
* @param values HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
*/
@JvmName("mgcxskqihjnoapxm")
public suspend fun archiveUris(vararg values: String) {
val toBeMapped = values.toList()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.archiveUris = mapped
}
/**
* @param value The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
*/
@JvmName("xrsukanllgfspkov")
public suspend fun args(`value`: List?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.args = mapped
}
/**
* @param values The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
*/
@JvmName("jteectjkfjjkguuh")
public suspend fun args(vararg values: String) {
val toBeMapped = values.toList()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.args = mapped
}
/**
* @param value HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
*/
@JvmName("dgceppmmlvcedacg")
public suspend fun fileUris(`value`: List?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.fileUris = mapped
}
/**
* @param values HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.
*/
@JvmName("eqcdgiyxardoajug")
public suspend fun fileUris(vararg values: String) {
val toBeMapped = values.toList()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.fileUris = mapped
}
/**
* @param value HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
*/
@JvmName("kxpssdfytgdcblmh")
public suspend fun jarFileUris(`value`: List?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.jarFileUris = mapped
}
/**
* @param values HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
*/
@JvmName("ygoexgsvjmtftknu")
public suspend fun jarFileUris(vararg values: String) {
val toBeMapped = values.toList()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.jarFileUris = mapped
}
/**
* @param value The runtime log config for job execution.
*/
@JvmName("ashmqtdjbhdioihu")
public suspend fun loggingConfig(`value`: WorkflowTemplateJobPysparkJobLoggingConfigArgs?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.loggingConfig = mapped
}
/**
* @param argument The runtime log config for job execution.
*/
@JvmName("ljbmqfettdrxxpcb")
public suspend fun loggingConfig(argument: suspend WorkflowTemplateJobPysparkJobLoggingConfigArgsBuilder.() -> Unit) {
val toBeMapped = WorkflowTemplateJobPysparkJobLoggingConfigArgsBuilder().applySuspend {
argument()
}.build()
val mapped = of(toBeMapped)
this.loggingConfig = mapped
}
/**
* @param value Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
*/
@JvmName("gyjfhagexugwbpab")
public suspend fun mainPythonFileUri(`value`: String) {
val toBeMapped = value
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.mainPythonFileUri = mapped
}
/**
* @param value A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
*/
@JvmName("fagtulfyxyfqkkny")
public suspend fun properties(`value`: Map?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.properties = mapped
}
/**
* @param values A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
*/
@JvmName("qmxhqipofmqjvnen")
public fun properties(vararg values: Pair) {
val toBeMapped = values.toMap()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.properties = mapped
}
/**
* @param value HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
*/
@JvmName("qfgoxkttnieehedu")
public suspend fun pythonFileUris(`value`: List?) {
val toBeMapped = value
val mapped = toBeMapped?.let({ args0 -> of(args0) })
this.pythonFileUris = mapped
}
/**
* @param values HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
*/
@JvmName("xypotkhhaxwjvuun")
public suspend fun pythonFileUris(vararg values: String) {
val toBeMapped = values.toList()
val mapped = toBeMapped.let({ args0 -> of(args0) })
this.pythonFileUris = mapped
}
internal fun build(): WorkflowTemplateJobPysparkJobArgs = WorkflowTemplateJobPysparkJobArgs(
archiveUris = archiveUris,
args = args,
fileUris = fileUris,
jarFileUris = jarFileUris,
loggingConfig = loggingConfig,
mainPythonFileUri = mainPythonFileUri ?: throw PulumiNullFieldException("mainPythonFileUri"),
properties = properties,
pythonFileUris = pythonFileUris,
)
}