Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
@file:Suppress("NAME_SHADOWING", "DEPRECATION")
package com.pulumi.awsnative.kinesisfirehose.kotlin.inputs
import com.pulumi.awsnative.kinesisfirehose.inputs.DeliveryStreamOrcSerDeArgs.builder
import com.pulumi.core.Output
import com.pulumi.core.Output.of
import com.pulumi.kotlin.ConvertibleToJava
import com.pulumi.kotlin.PulumiTagMarker
import kotlin.Boolean
import kotlin.Double
import kotlin.Int
import kotlin.String
import kotlin.Suppress
import kotlin.collections.List
import kotlin.jvm.JvmName
/**
*
* @property blockSizeBytes The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
* @property bloomFilterColumns The column names for which you want Firehose to create bloom filters. The default is `null` .
* @property bloomFilterFalsePositiveProbability The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
* @property compression The compression code to use over data blocks. The default is `SNAPPY` .
* @property dictionaryKeyThreshold Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
* @property enablePadding Set this to `true` to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is `false` .
* @property formatVersion The version of the file to write. The possible values are `V0_11` and `V0_12` . The default is `V0_12` .
* @property paddingTolerance A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
* For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
* Kinesis Data Firehose ignores this parameter when `EnablePadding` is `false` .
* @property rowIndexStride The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
* @property stripeSizeBytes The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
*/
public data class DeliveryStreamOrcSerDeArgs(
public val blockSizeBytes: Output? = null,
public val bloomFilterColumns: Output>? = null,
public val bloomFilterFalsePositiveProbability: Output? = null,
public val compression: Output? = null,
public val dictionaryKeyThreshold: Output? = null,
public val enablePadding: Output? = null,
public val formatVersion: Output? = null,
public val paddingTolerance: Output? = null,
public val rowIndexStride: Output? = null,
public val stripeSizeBytes: Output? = null,
) : ConvertibleToJava {
override fun toJava(): com.pulumi.awsnative.kinesisfirehose.inputs.DeliveryStreamOrcSerDeArgs =
com.pulumi.awsnative.kinesisfirehose.inputs.DeliveryStreamOrcSerDeArgs.builder()
.blockSizeBytes(blockSizeBytes?.applyValue({ args0 -> args0 }))
.bloomFilterColumns(bloomFilterColumns?.applyValue({ args0 -> args0.map({ args0 -> args0 }) }))
.bloomFilterFalsePositiveProbability(
bloomFilterFalsePositiveProbability?.applyValue({ args0 ->
args0
}),
)
.compression(compression?.applyValue({ args0 -> args0 }))
.dictionaryKeyThreshold(dictionaryKeyThreshold?.applyValue({ args0 -> args0 }))
.enablePadding(enablePadding?.applyValue({ args0 -> args0 }))
.formatVersion(formatVersion?.applyValue({ args0 -> args0 }))
.paddingTolerance(paddingTolerance?.applyValue({ args0 -> args0 }))
.rowIndexStride(rowIndexStride?.applyValue({ args0 -> args0 }))
.stripeSizeBytes(stripeSizeBytes?.applyValue({ args0 -> args0 })).build()
}
/**
* Builder for [DeliveryStreamOrcSerDeArgs].
*/
@PulumiTagMarker
public class DeliveryStreamOrcSerDeArgsBuilder internal constructor() {
private var blockSizeBytes: Output? = null
private var bloomFilterColumns: Output>? = null
private var bloomFilterFalsePositiveProbability: Output? = null
private var compression: Output? = null
private var dictionaryKeyThreshold: Output? = null
private var enablePadding: Output? = null
private var formatVersion: Output? = null
private var paddingTolerance: Output? = null
private var rowIndexStride: Output? = null
private var stripeSizeBytes: Output? = null
/**
* @param value The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
*/
@JvmName("rkwdahoecfnfjqyu")
public suspend fun blockSizeBytes(`value`: Output) {
this.blockSizeBytes = value
}
/**
* @param value The column names for which you want Firehose to create bloom filters. The default is `null` .
*/
@JvmName("ermuitsnnxmkwxva")
public suspend fun bloomFilterColumns(`value`: Output>) {
this.bloomFilterColumns = value
}
@JvmName("smtyneforpukiotc")
public suspend fun bloomFilterColumns(vararg values: Output) {
this.bloomFilterColumns = Output.all(values.asList())
}
/**
* @param values The column names for which you want Firehose to create bloom filters. The default is `null` .
*/
@JvmName("vdprdykjsubfncnf")
public suspend fun bloomFilterColumns(values: List