All Downloads are FREE. Search and download functionalities are using the official Maven repository.

com.pulumi.gcp.dataproc.kotlin.inputs.JobSparksqlConfigArgs.kt Maven / Gradle / Ivy

Go to download

Build cloud applications and infrastructure by combining the safety and reliability of infrastructure as code with the power of the Kotlin programming language.

There is a newer version: 8.10.0.0
Show newest version
@file:Suppress("NAME_SHADOWING", "DEPRECATION")

package com.pulumi.gcp.dataproc.kotlin.inputs

import com.pulumi.core.Output
import com.pulumi.core.Output.of
import com.pulumi.gcp.dataproc.inputs.JobSparksqlConfigArgs.builder
import com.pulumi.kotlin.ConvertibleToJava
import com.pulumi.kotlin.PulumiTagMarker
import com.pulumi.kotlin.applySuspend
import kotlin.Pair
import kotlin.String
import kotlin.Suppress
import kotlin.Unit
import kotlin.collections.List
import kotlin.collections.Map
import kotlin.jvm.JvmName

/**
 *
 * @property jarFileUris HCFS URIs of jar files to be added to the Spark CLASSPATH.
 * * `logging_config.driver_log_levels`- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
 * @property loggingConfig The runtime logging config of the job
 * @property properties A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
 * @property queryFileUri The HCFS URI of the script that contains SQL queries.
 * Conflicts with `query_list`
 * @property queryLists The list of SQL queries or statements to execute as part of the job.
 * Conflicts with `query_file_uri`
 * @property scriptVariables Mapping of query variable names to values (equivalent to the Spark SQL command: `SET name="value";`).
 */
public data class JobSparksqlConfigArgs(
    public val jarFileUris: Output>? = null,
    public val loggingConfig: Output? = null,
    public val properties: Output>? = null,
    public val queryFileUri: Output? = null,
    public val queryLists: Output>? = null,
    public val scriptVariables: Output>? = null,
) : ConvertibleToJava {
    override fun toJava(): com.pulumi.gcp.dataproc.inputs.JobSparksqlConfigArgs =
        com.pulumi.gcp.dataproc.inputs.JobSparksqlConfigArgs.builder()
            .jarFileUris(jarFileUris?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
            .loggingConfig(loggingConfig?.applyValue({ args0 -> args0.let({ args0 -> args0.toJava() }) }))
            .properties(
                properties?.applyValue({ args0 ->
                    args0.map({ args0 ->
                        args0.key.to(args0.value)
                    }).toMap()
                }),
            )
            .queryFileUri(queryFileUri?.applyValue({ args0 -> args0 }))
            .queryLists(queryLists?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
            .scriptVariables(
                scriptVariables?.applyValue({ args0 ->
                    args0.map({ args0 ->
                        args0.key.to(args0.value)
                    }).toMap()
                }),
            ).build()
}

/**
 * Builder for [JobSparksqlConfigArgs].
 */
@PulumiTagMarker
public class JobSparksqlConfigArgsBuilder internal constructor() {
    private var jarFileUris: Output>? = null

    private var loggingConfig: Output? = null

    private var properties: Output>? = null

    private var queryFileUri: Output? = null

    private var queryLists: Output>? = null

    private var scriptVariables: Output>? = null

    /**
     * @param value HCFS URIs of jar files to be added to the Spark CLASSPATH.
     * * `logging_config.driver_log_levels`- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
     */
    @JvmName("bkeeuayhdvjvsldo")
    public suspend fun jarFileUris(`value`: Output>) {
        this.jarFileUris = value
    }

    @JvmName("gxjipuvyurgauxua")
    public suspend fun jarFileUris(vararg values: Output) {
        this.jarFileUris = Output.all(values.asList())
    }

    /**
     * @param values HCFS URIs of jar files to be added to the Spark CLASSPATH.
     * * `logging_config.driver_log_levels`- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
     */
    @JvmName("flhekrupvpcgbmts")
    public suspend fun jarFileUris(values: List>) {
        this.jarFileUris = Output.all(values)
    }

    /**
     * @param value The runtime logging config of the job
     */
    @JvmName("akxbwgtupavqgyru")
    public suspend fun loggingConfig(`value`: Output) {
        this.loggingConfig = value
    }

    /**
     * @param value A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
     */
    @JvmName("lkjfmuhpmeeiymqb")
    public suspend fun properties(`value`: Output>) {
        this.properties = value
    }

    /**
     * @param value The HCFS URI of the script that contains SQL queries.
     * Conflicts with `query_list`
     */
    @JvmName("jmuefsvubogtbaja")
    public suspend fun queryFileUri(`value`: Output) {
        this.queryFileUri = value
    }

    /**
     * @param value The list of SQL queries or statements to execute as part of the job.
     * Conflicts with `query_file_uri`
     */
    @JvmName("ttxgucrjxixxleij")
    public suspend fun queryLists(`value`: Output>) {
        this.queryLists = value
    }

    @JvmName("bnuurpfxjhthiuen")
    public suspend fun queryLists(vararg values: Output) {
        this.queryLists = Output.all(values.asList())
    }

    /**
     * @param values The list of SQL queries or statements to execute as part of the job.
     * Conflicts with `query_file_uri`
     */
    @JvmName("aeewmpsqhwiwgfpm")
    public suspend fun queryLists(values: List>) {
        this.queryLists = Output.all(values)
    }

    /**
     * @param value Mapping of query variable names to values (equivalent to the Spark SQL command: `SET name="value";`).
     */
    @JvmName("mtnrpgrnhwqwvqwt")
    public suspend fun scriptVariables(`value`: Output>) {
        this.scriptVariables = value
    }

    /**
     * @param value HCFS URIs of jar files to be added to the Spark CLASSPATH.
     * * `logging_config.driver_log_levels`- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
     */
    @JvmName("bifoknqlflembaow")
    public suspend fun jarFileUris(`value`: List?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.jarFileUris = mapped
    }

    /**
     * @param values HCFS URIs of jar files to be added to the Spark CLASSPATH.
     * * `logging_config.driver_log_levels`- (Required) The per-package log levels for the driver. This may include 'root' package name to configure rootLogger. Examples: 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG'
     */
    @JvmName("hhehpkcxoyqochls")
    public suspend fun jarFileUris(vararg values: String) {
        val toBeMapped = values.toList()
        val mapped = toBeMapped.let({ args0 -> of(args0) })
        this.jarFileUris = mapped
    }

    /**
     * @param value The runtime logging config of the job
     */
    @JvmName("kivylxgxrbuqlfgn")
    public suspend fun loggingConfig(`value`: JobSparksqlConfigLoggingConfigArgs?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.loggingConfig = mapped
    }

    /**
     * @param argument The runtime logging config of the job
     */
    @JvmName("umgwihmgftjiofxh")
    public suspend fun loggingConfig(argument: suspend JobSparksqlConfigLoggingConfigArgsBuilder.() -> Unit) {
        val toBeMapped = JobSparksqlConfigLoggingConfigArgsBuilder().applySuspend { argument() }.build()
        val mapped = of(toBeMapped)
        this.loggingConfig = mapped
    }

    /**
     * @param value A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
     */
    @JvmName("jhtrqmcfgvystwgr")
    public suspend fun properties(`value`: Map?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.properties = mapped
    }

    /**
     * @param values A mapping of property names to values, used to configure Spark SQL's SparkConf. Properties that conflict with values set by the Cloud Dataproc API may be overwritten.
     */
    @JvmName("qhvkcqqxrjsdsxxs")
    public fun properties(vararg values: Pair) {
        val toBeMapped = values.toMap()
        val mapped = toBeMapped.let({ args0 -> of(args0) })
        this.properties = mapped
    }

    /**
     * @param value The HCFS URI of the script that contains SQL queries.
     * Conflicts with `query_list`
     */
    @JvmName("nmfqrgktwfbnxvru")
    public suspend fun queryFileUri(`value`: String?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.queryFileUri = mapped
    }

    /**
     * @param value The list of SQL queries or statements to execute as part of the job.
     * Conflicts with `query_file_uri`
     */
    @JvmName("nwrhtxuovqengiby")
    public suspend fun queryLists(`value`: List?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.queryLists = mapped
    }

    /**
     * @param values The list of SQL queries or statements to execute as part of the job.
     * Conflicts with `query_file_uri`
     */
    @JvmName("treaqyrrdghogkvp")
    public suspend fun queryLists(vararg values: String) {
        val toBeMapped = values.toList()
        val mapped = toBeMapped.let({ args0 -> of(args0) })
        this.queryLists = mapped
    }

    /**
     * @param value Mapping of query variable names to values (equivalent to the Spark SQL command: `SET name="value";`).
     */
    @JvmName("hpjaadkcdbwjuulr")
    public suspend fun scriptVariables(`value`: Map?) {
        val toBeMapped = value
        val mapped = toBeMapped?.let({ args0 -> of(args0) })
        this.scriptVariables = mapped
    }

    /**
     * @param values Mapping of query variable names to values (equivalent to the Spark SQL command: `SET name="value";`).
     */
    @JvmName("qpbjunlovgybydhx")
    public fun scriptVariables(vararg values: Pair) {
        val toBeMapped = values.toMap()
        val mapped = toBeMapped.let({ args0 -> of(args0) })
        this.scriptVariables = mapped
    }

    internal fun build(): JobSparksqlConfigArgs = JobSparksqlConfigArgs(
        jarFileUris = jarFileUris,
        loggingConfig = loggingConfig,
        properties = properties,
        queryFileUri = queryFileUri,
        queryLists = queryLists,
        scriptVariables = scriptVariables,
    )
}




© 2015 - 2024 Weber Informatics LLC | Privacy Policy