We recommend new projects start with resources from the AWS provider.
aws-native.kinesisfirehose.getDeliveryStream
Explore with Pulumi AI
We recommend new projects start with resources from the AWS provider.
Resource Type definition for AWS::KinesisFirehose::DeliveryStream
Using getDeliveryStream
Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.
function getDeliveryStream(args: GetDeliveryStreamArgs, opts?: InvokeOptions): Promise<GetDeliveryStreamResult>
function getDeliveryStreamOutput(args: GetDeliveryStreamOutputArgs, opts?: InvokeOptions): Output<GetDeliveryStreamResult>
def get_delivery_stream(delivery_stream_name: Optional[str] = None,
opts: Optional[InvokeOptions] = None) -> GetDeliveryStreamResult
def get_delivery_stream_output(delivery_stream_name: Optional[pulumi.Input[str]] = None,
opts: Optional[InvokeOptions] = None) -> Output[GetDeliveryStreamResult]
func LookupDeliveryStream(ctx *Context, args *LookupDeliveryStreamArgs, opts ...InvokeOption) (*LookupDeliveryStreamResult, error)
func LookupDeliveryStreamOutput(ctx *Context, args *LookupDeliveryStreamOutputArgs, opts ...InvokeOption) LookupDeliveryStreamResultOutput
> Note: This function is named LookupDeliveryStream
in the Go SDK.
public static class GetDeliveryStream
{
public static Task<GetDeliveryStreamResult> InvokeAsync(GetDeliveryStreamArgs args, InvokeOptions? opts = null)
public static Output<GetDeliveryStreamResult> Invoke(GetDeliveryStreamInvokeArgs args, InvokeOptions? opts = null)
}
public static CompletableFuture<GetDeliveryStreamResult> getDeliveryStream(GetDeliveryStreamArgs args, InvokeOptions options)
public static Output<GetDeliveryStreamResult> getDeliveryStream(GetDeliveryStreamArgs args, InvokeOptions options)
fn::invoke:
function: aws-native:kinesisfirehose:getDeliveryStream
arguments:
# arguments dictionary
The following arguments are supported:
- Delivery
Stream stringName - The name of the Firehose stream.
- Delivery
Stream stringName - The name of the Firehose stream.
- delivery
Stream StringName - The name of the Firehose stream.
- delivery
Stream stringName - The name of the Firehose stream.
- delivery_
stream_ strname - The name of the Firehose stream.
- delivery
Stream StringName - The name of the Firehose stream.
getDeliveryStream Result
The following output properties are available:
- Arn string
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - Delivery
Stream Pulumi.Encryption Configuration Input Aws Native. Kinesis Firehose. Outputs. Delivery Stream Encryption Configuration Input - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- Extended
S3Destination Pulumi.Configuration Aws Native. Kinesis Firehose. Outputs. Delivery Stream Extended S3Destination Configuration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- Http
Endpoint Pulumi.Destination Configuration Aws Native. Kinesis Firehose. Outputs. Delivery Stream Http Endpoint Destination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- Redshift
Destination Pulumi.Configuration Aws Native. Kinesis Firehose. Outputs. Delivery Stream Redshift Destination Configuration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- Splunk
Destination Pulumi.Configuration Aws Native. Kinesis Firehose. Outputs. Delivery Stream Splunk Destination Configuration - The configuration of a destination in Splunk for the delivery stream.
- List<Pulumi.
Aws Native. Outputs. Tag> A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
- Arn string
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - Delivery
Stream DeliveryEncryption Configuration Input Stream Encryption Configuration Input Type - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- Extended
S3Destination DeliveryConfiguration Stream Extended S3Destination Configuration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- Http
Endpoint DeliveryDestination Configuration Stream Http Endpoint Destination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- Redshift
Destination DeliveryConfiguration Stream Redshift Destination Configuration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- Splunk
Destination DeliveryConfiguration Stream Splunk Destination Configuration - The configuration of a destination in Splunk for the delivery stream.
- Tag
A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
- arn String
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - delivery
Stream DeliveryEncryption Configuration Input Stream Encryption Configuration Input - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- extended
S3Destination DeliveryConfiguration Stream Extended S3Destination Configuration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- http
Endpoint DeliveryDestination Configuration Stream Http Endpoint Destination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- redshift
Destination DeliveryConfiguration Stream Redshift Destination Configuration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- splunk
Destination DeliveryConfiguration Stream Splunk Destination Configuration - The configuration of a destination in Splunk for the delivery stream.
- List<Tag>
A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
- arn string
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - delivery
Stream DeliveryEncryption Configuration Input Stream Encryption Configuration Input - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- extended
S3Destination DeliveryConfiguration Stream Extended S3Destination Configuration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- http
Endpoint DeliveryDestination Configuration Stream Http Endpoint Destination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- redshift
Destination DeliveryConfiguration Stream Redshift Destination Configuration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- splunk
Destination DeliveryConfiguration Stream Splunk Destination Configuration - The configuration of a destination in Splunk for the delivery stream.
- Tag[]
A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
- arn str
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - delivery_
stream_ Deliveryencryption_ configuration_ input Stream Encryption Configuration Input - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- extended_
s3_ Deliverydestination_ configuration Stream Extended S3Destination Configuration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- http_
endpoint_ Deliverydestination_ configuration Stream Http Endpoint Destination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- redshift_
destination_ Deliveryconfiguration Stream Redshift Destination Configuration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- splunk_
destination_ Deliveryconfiguration Stream Splunk Destination Configuration - The configuration of a destination in Splunk for the delivery stream.
- Sequence[root_Tag]
A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
- arn String
- The Amazon Resource Name (ARN) of the delivery stream, such as
arn:aws:firehose:us-east-2:123456789012:deliverystream/delivery-stream-name
. - delivery
Stream Property MapEncryption Configuration Input - Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side Encryption (SSE).
- extended
S3Destination Property MapConfiguration An Amazon S3 destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Extended S3 destination to an Amazon ES destination, update requires some interruptions .
- http
Endpoint Property MapDestination Configuration - Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination. You can specify only one destination.
- redshift
Destination Property MapConfiguration An Amazon Redshift destination for the delivery stream.
Conditional. You must specify only one destination configuration.
If you change the delivery stream destination from an Amazon Redshift destination to an Amazon ES destination, update requires some interruptions .
- splunk
Destination Property MapConfiguration - The configuration of a destination in Splunk for the delivery stream.
- List<Property Map>
A set of tags to assign to the Firehose stream. A tag is a key-value pair that you can define and assign to AWS resources. Tags are metadata. For example, you can add friendly names and descriptions or other types of information that can help you distinguish the Firehose stream. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.
You can specify up to 50 tags when creating a Firehose stream.
If you specify tags in the
CreateDeliveryStream
action, Amazon Data Firehose performs an additional authorization on thefirehose:TagDeliveryStream
action to verify if users have permissions to create tags. If you do not provide this permission, requests to create new Firehose streams with IAM resource tags will fail with anAccessDeniedException
such as following.AccessDeniedException
User: arn:aws:sts::x:assumed-role/x/x is not authorized to perform: firehose:TagDeliveryStream on resource: arn:aws:firehose:us-east-1:x:deliverystream/x with an explicit deny in an identity-based policy.
For an example IAM policy, see Tag example.
Supporting Types
DeliveryStreamBufferingHints
- Interval
In intSeconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - Size
In intMbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
- Interval
In intSeconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - Size
In intMbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
- interval
In IntegerSeconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - size
In IntegerMbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
- interval
In numberSeconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - size
In numberMbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
- interval_
in_ intseconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - size_
in_ intmbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
- interval
In NumberSeconds - The length of time, in seconds, that Kinesis Data Firehose buffers incoming data before delivering it to the destination. For valid values, see the
IntervalInSeconds
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference . - size
In NumberMbs - The size of the buffer, in MBs, that Kinesis Data Firehose uses for incoming data before delivering it to the destination. For valid values, see the
SizeInMBs
content for the BufferingHints data type in the Amazon Kinesis Data Firehose API Reference .
DeliveryStreamCloudWatchLoggingOptions
- Enabled bool
- Indicates whether CloudWatch Logs logging is enabled.
- Log
Group stringName The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- Log
Stream stringName The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
- Enabled bool
- Indicates whether CloudWatch Logs logging is enabled.
- Log
Group stringName The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- Log
Stream stringName The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
- enabled Boolean
- Indicates whether CloudWatch Logs logging is enabled.
- log
Group StringName The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- log
Stream StringName The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
- enabled boolean
- Indicates whether CloudWatch Logs logging is enabled.
- log
Group stringName The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- log
Stream stringName The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
- enabled bool
- Indicates whether CloudWatch Logs logging is enabled.
- log_
group_ strname The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- log_
stream_ strname The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
- enabled Boolean
- Indicates whether CloudWatch Logs logging is enabled.
- log
Group StringName The name of the CloudWatch Logs log group that contains the log stream that Kinesis Data Firehose will use.
Conditional. If you enable logging, you must specify this property.
- log
Stream StringName The name of the CloudWatch Logs log stream that Kinesis Data Firehose uses to send logs about data delivery.
Conditional. If you enable logging, you must specify this property.
DeliveryStreamCopyCommand
- Data
Table stringName - The name of the target table. The table must already exist in the database.
- Copy
Options string - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - Data
Table stringColumns - A comma-separated list of column names.
- Data
Table stringName - The name of the target table. The table must already exist in the database.
- Copy
Options string - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - Data
Table stringColumns - A comma-separated list of column names.
- data
Table StringName - The name of the target table. The table must already exist in the database.
- copy
Options String - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - data
Table StringColumns - A comma-separated list of column names.
- data
Table stringName - The name of the target table. The table must already exist in the database.
- copy
Options string - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - data
Table stringColumns - A comma-separated list of column names.
- data_
table_ strname - The name of the target table. The table must already exist in the database.
- copy_
options str - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - data_
table_ strcolumns - A comma-separated list of column names.
- data
Table StringName - The name of the target table. The table must already exist in the database.
- copy
Options String - Parameters to use with the Amazon Redshift
COPY
command. For examples, see theCopyOptions
content for the CopyCommand data type in the Amazon Kinesis Data Firehose API Reference . - data
Table StringColumns - A comma-separated list of column names.
DeliveryStreamDataFormatConversionConfiguration
- Enabled bool
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - Input
Format Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Input Format Configuration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - Output
Format Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Output Format Configuration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - Schema
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Schema Configuration - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
- Enabled bool
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - Input
Format DeliveryConfiguration Stream Input Format Configuration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - Output
Format DeliveryConfiguration Stream Output Format Configuration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - Schema
Configuration DeliveryStream Schema Configuration - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
- enabled Boolean
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - input
Format DeliveryConfiguration Stream Input Format Configuration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - output
Format DeliveryConfiguration Stream Output Format Configuration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - schema
Configuration DeliveryStream Schema Configuration - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
- enabled boolean
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - input
Format DeliveryConfiguration Stream Input Format Configuration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - output
Format DeliveryConfiguration Stream Output Format Configuration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - schema
Configuration DeliveryStream Schema Configuration - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
- enabled bool
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - input_
format_ Deliveryconfiguration Stream Input Format Configuration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - output_
format_ Deliveryconfiguration Stream Output Format Configuration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - schema_
configuration DeliveryStream Schema Configuration - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
- enabled Boolean
- Defaults to
true
. Set it tofalse
if you want to disable format conversion while preserving the configuration details. - input
Format Property MapConfiguration - Specifies the deserializer that you want Firehose to use to convert the format of your data from JSON. This parameter is required if
Enabled
is set to true. - output
Format Property MapConfiguration - Specifies the serializer that you want Firehose to use to convert the format of your data to the Parquet or ORC format. This parameter is required if
Enabled
is set to true. - schema
Configuration Property Map - Specifies the AWS Glue Data Catalog table that contains the column information. This parameter is required if
Enabled
is set to true.
DeliveryStreamDeserializer
- Hive
Json Pulumi.Ser De Aws Native. Kinesis Firehose. Inputs. Delivery Stream Hive Json Ser De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- Open
XJson Pulumi.Ser De Aws Native. Kinesis Firehose. Inputs. Delivery Stream Open XJson Ser De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
- Hive
Json DeliverySer De Stream Hive Json Ser De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- Open
XJson DeliverySer De Stream Open XJson Ser De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
- hive
Json DeliverySer De Stream Hive Json Ser De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- open
XJson DeliverySer De Stream Open XJson Ser De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
- hive
Json DeliverySer De Stream Hive Json Ser De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- open
XJson DeliverySer De Stream Open XJson Ser De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
- hive_
json_ Deliveryser_ de Stream Hive Json Ser De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- open_
x_ Deliveryjson_ ser_ de Stream Open XJson Ser De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
- hive
Json Property MapSer De - The native Hive / HCatalog JsonSerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the OpenX SerDe.
- open
XJson Property MapSer De - The OpenX SerDe. Used by Firehose for deserializing data, which means converting it from the JSON format in preparation for serializing it to the Parquet or ORC format. This is one of two deserializers you can choose, depending on which one offers the functionality you need. The other option is the native Hive / HCatalog JsonSerDe.
DeliveryStreamDynamicPartitioningConfiguration
- Enabled bool
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- Retry
Options Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Retry Options - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
- Enabled bool
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- Retry
Options DeliveryStream Retry Options - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
- enabled Boolean
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- retry
Options DeliveryStream Retry Options - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
- enabled boolean
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- retry
Options DeliveryStream Retry Options - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
- enabled bool
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- retry_
options DeliveryStream Retry Options - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
- enabled Boolean
- Specifies whether dynamic partitioning is enabled for this Kinesis Data Firehose delivery stream.
- retry
Options Property Map - Specifies the retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon S3 prefix.
DeliveryStreamEncryptionConfiguration
- Kms
Encryption Pulumi.Config Aws Native. Kinesis Firehose. Inputs. Delivery Stream Kms Encryption Config - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- No
Encryption Pulumi.Config Aws Native. Kinesis Firehose. Delivery Stream Encryption Configuration No Encryption Config - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
- Kms
Encryption DeliveryConfig Stream Kms Encryption Config - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- No
Encryption DeliveryConfig Stream Encryption Configuration No Encryption Config - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
- kms
Encryption DeliveryConfig Stream Kms Encryption Config - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- no
Encryption DeliveryConfig Stream Encryption Configuration No Encryption Config - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
- kms
Encryption DeliveryConfig Stream Kms Encryption Config - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- no
Encryption DeliveryConfig Stream Encryption Configuration No Encryption Config - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
- kms_
encryption_ Deliveryconfig Stream Kms Encryption Config - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- no_
encryption_ Deliveryconfig Stream Encryption Configuration No Encryption Config - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
- kms
Encryption Property MapConfig - The AWS Key Management Service ( AWS KMS) encryption key that Amazon S3 uses to encrypt your data.
- no
Encryption "NoConfig Encryption" - Disables encryption. For valid values, see the
NoEncryptionConfig
content for the EncryptionConfiguration data type in the Amazon Kinesis Data Firehose API Reference .
DeliveryStreamEncryptionConfigurationInput
- Key
Type Pulumi.Aws Native. Kinesis Firehose. Delivery Stream Encryption Configuration Input Key Type Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- Key
Arn string - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
- Key
Type DeliveryStream Encryption Configuration Input Key Type Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- Key
Arn string - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
- key
Type DeliveryStream Encryption Configuration Input Key Type Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- key
Arn String - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
- key
Type DeliveryStream Encryption Configuration Input Key Type Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- key
Arn string - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
- key_
type DeliveryStream Encryption Configuration Input Key Type Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- key_
arn str - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
- key
Type "AWS_OWNED_CMK" | "CUSTOMER_MANAGED_CMK" Indicates the type of customer master key (CMK) to use for encryption. The default setting is
AWS_OWNED_CMK
. For more information about CMKs, see Customer Master Keys (CMKs) .You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams.
To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see About Symmetric and Asymmetric CMKs in the AWS Key Management Service developer guide.
- key
Arn String - If you set
KeyType
toCUSTOMER_MANAGED_CMK
, you must specify the Amazon Resource Name (ARN) of the CMK. If you setKeyType
toAWS _OWNED_CMK
, Firehose uses a service-account CMK.
DeliveryStreamEncryptionConfigurationInputKeyType
DeliveryStreamEncryptionConfigurationNoEncryptionConfig
DeliveryStreamExtendedS3DestinationConfiguration
- Bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- Role
Arn string - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- Buffering
Hints Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Buffering Hints - The buffering option.
- Cloud
Watch Pulumi.Logging Options Aws Native. Kinesis Firehose. Inputs. Delivery Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- Compression
Format Pulumi.Aws Native. Kinesis Firehose. Delivery Stream Extended S3Destination Configuration Compression Format - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - Custom
Time stringZone - The time zone you prefer. UTC is the default.
- Data
Format Pulumi.Conversion Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Data Format Conversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- Dynamic
Partitioning Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Dynamic Partitioning Configuration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- Encryption
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Encryption Configuration - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - Error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- File
Extension string - Specify a file extension. It will override the default file extension
- Prefix string
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - Processing
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- S3Backup
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream S3Destination Configuration - The configuration for backup in Amazon S3.
- S3Backup
Mode Pulumi.Aws Native. Kinesis Firehose. Delivery Stream Extended S3Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- Bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- Role
Arn string - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- Buffering
Hints DeliveryStream Buffering Hints - The buffering option.
- Cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- Compression
Format DeliveryStream Extended S3Destination Configuration Compression Format - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - Custom
Time stringZone - The time zone you prefer. UTC is the default.
- Data
Format DeliveryConversion Configuration Stream Data Format Conversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- Dynamic
Partitioning DeliveryConfiguration Stream Dynamic Partitioning Configuration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- Encryption
Configuration DeliveryStream Encryption Configuration - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - Error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- File
Extension string - Specify a file extension. It will override the default file extension
- Prefix string
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - Processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- S3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- S3Backup
Mode DeliveryStream Extended S3Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- bucket
Arn String - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- role
Arn String - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- buffering
Hints DeliveryStream Buffering Hints - The buffering option.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- compression
Format DeliveryStream Extended S3Destination Configuration Compression Format - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - custom
Time StringZone - The time zone you prefer. UTC is the default.
- data
Format DeliveryConversion Configuration Stream Data Format Conversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- dynamic
Partitioning DeliveryConfiguration Stream Dynamic Partitioning Configuration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- encryption
Configuration DeliveryStream Encryption Configuration - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - error
Output StringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- file
Extension String - Specify a file extension. It will override the default file extension
- prefix String
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- s3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3Backup
Mode DeliveryStream Extended S3Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- role
Arn string - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- buffering
Hints DeliveryStream Buffering Hints - The buffering option.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- compression
Format DeliveryStream Extended S3Destination Configuration Compression Format - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - custom
Time stringZone - The time zone you prefer. UTC is the default.
- data
Format DeliveryConversion Configuration Stream Data Format Conversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- dynamic
Partitioning DeliveryConfiguration Stream Dynamic Partitioning Configuration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- encryption
Configuration DeliveryStream Encryption Configuration - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- file
Extension string - Specify a file extension. It will override the default file extension
- prefix string
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- s3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3Backup
Mode DeliveryStream Extended S3Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- bucket_
arn str - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- role_
arn str - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- buffering_
hints DeliveryStream Buffering Hints - The buffering option.
- cloud_
watch_ Deliverylogging_ options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- compression_
format DeliveryStream Extended S3Destination Configuration Compression Format - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - custom_
time_ strzone - The time zone you prefer. UTC is the default.
- data_
format_ Deliveryconversion_ configuration Stream Data Format Conversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- dynamic_
partitioning_ Deliveryconfiguration Stream Dynamic Partitioning Configuration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- encryption_
configuration DeliveryStream Encryption Configuration - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - error_
output_ strprefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- file_
extension str - Specify a file extension. It will override the default file extension
- prefix str
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - processing_
configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- s3_
backup_ Deliveryconfiguration Stream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3_
backup_ Deliverymode Stream Extended S3Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- bucket
Arn String - The Amazon Resource Name (ARN) of the Amazon S3 bucket. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- role
Arn String - The Amazon Resource Name (ARN) of the AWS credentials. For constraints, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference .
- buffering
Hints Property Map - The buffering option.
- cloud
Watch Property MapLogging Options - The Amazon CloudWatch logging options for your Firehose stream.
- compression
Format "UNCOMPRESSED" | "GZIP" | "ZIP" | "Snappy" | "HADOOP_SNAPPY" - The compression format. If no value is specified, the default is
UNCOMPRESSED
. - custom
Time StringZone - The time zone you prefer. UTC is the default.
- data
Format Property MapConversion Configuration - The serializer, deserializer, and schema for converting data from the JSON format to the Parquet or ORC format before writing it to Amazon S3.
- dynamic
Partitioning Property MapConfiguration - The configuration of the dynamic partitioning mechanism that creates targeted data sets from the streaming data by partitioning it based on partition keys.
- encryption
Configuration Property Map - The encryption configuration for the Kinesis Data Firehose delivery stream. The default value is
NoEncryption
. - error
Output StringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- file
Extension String - Specify a file extension. It will override the default file extension
- prefix String
- The
YYYY/MM/DD/HH
time format prefix is automatically used for delivered Amazon S3 files. For more information, see ExtendedS3DestinationConfiguration in the Amazon Kinesis Data Firehose API Reference . - processing
Configuration Property Map - The data processing configuration for the Kinesis Data Firehose delivery stream.
- s3Backup
Configuration Property Map - The configuration for backup in Amazon S3.
- s3Backup
Mode "Disabled" | "Enabled" - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
DeliveryStreamExtendedS3DestinationConfigurationCompressionFormat
DeliveryStreamExtendedS3DestinationConfigurationS3BackupMode
DeliveryStreamHiveJsonSerDe
- Timestamp
Formats List<string> - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
- Timestamp
Formats []string - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
- timestamp
Formats List<String> - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
- timestamp
Formats string[] - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
- timestamp_
formats Sequence[str] - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
- timestamp
Formats List<String> - Indicates how you want Firehose to parse the date and timestamps that may be present in your input data JSON. To specify these format strings, follow the pattern syntax of JodaTime's DateTimeFormat format strings. For more information, see Class DateTimeFormat . You can also use the special value
millis
to parse timestamps in epoch milliseconds. If you don't specify a format, Firehose usesjava.sql.Timestamp::valueOf
by default.
DeliveryStreamHttpEndpointCommonAttribute
- Attribute
Name string - The name of the HTTP endpoint common attribute.
- Attribute
Value string - The value of the HTTP endpoint common attribute.
- Attribute
Name string - The name of the HTTP endpoint common attribute.
- Attribute
Value string - The value of the HTTP endpoint common attribute.
- attribute
Name String - The name of the HTTP endpoint common attribute.
- attribute
Value String - The value of the HTTP endpoint common attribute.
- attribute
Name string - The name of the HTTP endpoint common attribute.
- attribute
Value string - The value of the HTTP endpoint common attribute.
- attribute_
name str - The name of the HTTP endpoint common attribute.
- attribute_
value str - The value of the HTTP endpoint common attribute.
- attribute
Name String - The name of the HTTP endpoint common attribute.
- attribute
Value String - The value of the HTTP endpoint common attribute.
DeliveryStreamHttpEndpointConfiguration
- url str
- The URL of the HTTP endpoint selected as the destination.
- access_
key str - The access key required for Kinesis Firehose to authenticate with the HTTP endpoint selected as the destination.
- name str
- The name of the HTTP endpoint selected as the destination.
DeliveryStreamHttpEndpointDestinationConfiguration
- Endpoint
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Http Endpoint Configuration - The configuration of the HTTP endpoint selected as the destination.
- S3Configuration
Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream S3Destination Configuration - Describes the configuration of a destination in Amazon S3.
- Buffering
Hints Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Buffering Hints - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- Cloud
Watch Pulumi.Logging Options Aws Native. Kinesis Firehose. Inputs. Delivery Stream Cloud Watch Logging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- Processing
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processing Configuration - Describes the data processing configuration.
- Request
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Http Endpoint Request Configuration - The configuration of the request sent to the HTTP endpoint specified as the destination.
- Retry
Options Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Retry Options - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- Role
Arn string - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- S3Backup
Mode string - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- Secrets
Manager Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Secrets Manager Configuration - The configuration that defines how you access secrets for HTTP Endpoint destination.
- Endpoint
Configuration DeliveryStream Http Endpoint Configuration - The configuration of the HTTP endpoint selected as the destination.
- S3Configuration
Delivery
Stream S3Destination Configuration - Describes the configuration of a destination in Amazon S3.
- Buffering
Hints DeliveryStream Buffering Hints - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- Cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- Processing
Configuration DeliveryStream Processing Configuration - Describes the data processing configuration.
- Request
Configuration DeliveryStream Http Endpoint Request Configuration - The configuration of the request sent to the HTTP endpoint specified as the destination.
- Retry
Options DeliveryStream Retry Options - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- Role
Arn string - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- S3Backup
Mode string - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- Secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for HTTP Endpoint destination.
- endpoint
Configuration DeliveryStream Http Endpoint Configuration - The configuration of the HTTP endpoint selected as the destination.
- s3Configuration
Delivery
Stream S3Destination Configuration - Describes the configuration of a destination in Amazon S3.
- buffering
Hints DeliveryStream Buffering Hints - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- processing
Configuration DeliveryStream Processing Configuration - Describes the data processing configuration.
- request
Configuration DeliveryStream Http Endpoint Request Configuration - The configuration of the request sent to the HTTP endpoint specified as the destination.
- retry
Options DeliveryStream Retry Options - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- role
Arn String - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- s3Backup
Mode String - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for HTTP Endpoint destination.
- endpoint
Configuration DeliveryStream Http Endpoint Configuration - The configuration of the HTTP endpoint selected as the destination.
- s3Configuration
Delivery
Stream S3Destination Configuration - Describes the configuration of a destination in Amazon S3.
- buffering
Hints DeliveryStream Buffering Hints - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- processing
Configuration DeliveryStream Processing Configuration - Describes the data processing configuration.
- request
Configuration DeliveryStream Http Endpoint Request Configuration - The configuration of the request sent to the HTTP endpoint specified as the destination.
- retry
Options DeliveryStream Retry Options - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- role
Arn string - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- s3Backup
Mode string - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for HTTP Endpoint destination.
- endpoint_
configuration DeliveryStream Http Endpoint Configuration - The configuration of the HTTP endpoint selected as the destination.
- s3_
configuration DeliveryStream S3Destination Configuration - Describes the configuration of a destination in Amazon S3.
- buffering_
hints DeliveryStream Buffering Hints - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- cloud_
watch_ Deliverylogging_ options Stream Cloud Watch Logging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- processing_
configuration DeliveryStream Processing Configuration - Describes the data processing configuration.
- request_
configuration DeliveryStream Http Endpoint Request Configuration - The configuration of the request sent to the HTTP endpoint specified as the destination.
- retry_
options DeliveryStream Retry Options - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- role_
arn str - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- s3_
backup_ strmode - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- secrets_
manager_ Deliveryconfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for HTTP Endpoint destination.
- endpoint
Configuration Property Map - The configuration of the HTTP endpoint selected as the destination.
- s3Configuration Property Map
- Describes the configuration of a destination in Amazon S3.
- buffering
Hints Property Map - The buffering options that can be used before data is delivered to the specified destination. Kinesis Data Firehose treats these options as hints, and it might choose to use more optimal values. The SizeInMBs and IntervalInSeconds parameters are optional. However, if you specify a value for one of them, you must also provide a value for the other.
- cloud
Watch Property MapLogging Options - Describes the Amazon CloudWatch logging options for your delivery stream.
- processing
Configuration Property Map - Describes the data processing configuration.
- request
Configuration Property Map - The configuration of the request sent to the HTTP endpoint specified as the destination.
- retry
Options Property Map - Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of receipt from the specified HTTP endpoint destination.
- role
Arn String - Kinesis Data Firehose uses this IAM role for all the permissions that the delivery stream needs.
- s3Backup
Mode String - Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers to the HTTP endpoint destination. You can back up all documents (AllData) or only the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint destination (FailedDataOnly).
- secrets
Manager Property MapConfiguration - The configuration that defines how you access secrets for HTTP Endpoint destination.
DeliveryStreamHttpEndpointRequestConfiguration
- Common
Attributes List<Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Http Endpoint Common Attribute> - Describes the metadata sent to the HTTP endpoint destination.
- Content
Encoding Pulumi.Aws Native. Kinesis Firehose. Delivery Stream Http Endpoint Request Configuration Content Encoding - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
- Common
Attributes []DeliveryStream Http Endpoint Common Attribute - Describes the metadata sent to the HTTP endpoint destination.
- Content
Encoding DeliveryStream Http Endpoint Request Configuration Content Encoding - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
- common
Attributes List<DeliveryStream Http Endpoint Common Attribute> - Describes the metadata sent to the HTTP endpoint destination.
- content
Encoding DeliveryStream Http Endpoint Request Configuration Content Encoding - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
- common
Attributes DeliveryStream Http Endpoint Common Attribute[] - Describes the metadata sent to the HTTP endpoint destination.
- content
Encoding DeliveryStream Http Endpoint Request Configuration Content Encoding - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
- common_
attributes Sequence[DeliveryStream Http Endpoint Common Attribute] - Describes the metadata sent to the HTTP endpoint destination.
- content_
encoding DeliveryStream Http Endpoint Request Configuration Content Encoding - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
- common
Attributes List<Property Map> - Describes the metadata sent to the HTTP endpoint destination.
- content
Encoding "NONE" | "GZIP" - Kinesis Data Firehose uses the content encoding to compress the body of a request before sending the request to the destination. For more information, see Content-Encoding in MDN Web Docs, the official Mozilla documentation.
DeliveryStreamHttpEndpointRequestConfigurationContentEncoding
DeliveryStreamInputFormatConfiguration
- Deserializer
Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream Deserializer - Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
- Deserializer
Delivery
Stream Deserializer - Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
- deserializer
Delivery
Stream Deserializer - Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
- deserializer
Delivery
Stream Deserializer - Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
- deserializer
Delivery
Stream Deserializer - Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
- deserializer Property Map
- Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe or the OpenX JSON SerDe. If both are non-null, the server rejects the request.
DeliveryStreamKmsEncryptionConfig
- Awskms
Key stringArn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
- Awskms
Key stringArn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
- awskms
Key StringArn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
- awskms
Key stringArn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
- awskms_
key_ strarn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
- awskms
Key StringArn - The Amazon Resource Name (ARN) of the AWS KMS encryption key that Amazon S3 uses to encrypt data delivered by the Kinesis Data Firehose stream. The key must belong to the same region as the destination S3 bucket.
DeliveryStreamOpenXJsonSerDe
- Case
Insensitive bool - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - Column
To Dictionary<string, string>Json Key Mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - Convert
Dots boolIn Json Keys To Underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
- Case
Insensitive bool - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - Column
To map[string]stringJson Key Mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - Convert
Dots boolIn Json Keys To Underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
- case
Insensitive Boolean - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - column
To Map<String,String>Json Key Mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - convert
Dots BooleanIn Json Keys To Underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
- case
Insensitive boolean - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - column
To {[key: string]: string}Json Key Mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - convert
Dots booleanIn Json Keys To Underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
- case_
insensitive bool - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - column_
to_ Mapping[str, str]json_ key_ mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - convert_
dots_ boolin_ json_ keys_ to_ underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
- case
Insensitive Boolean - When set to
true
, which is the default, Firehose converts JSON keys to lowercase before deserializing them. - column
To Map<String>Json Key Mappings - Maps column names to JSON keys that aren't identical to the column names. This is useful when the JSON contains keys that are Hive keywords. For example,
timestamp
is a Hive keyword. If you have a JSON key namedtimestamp
, set this parameter to{"ts": "timestamp"}
to map this key to a column namedts
. - convert
Dots BooleanIn Json Keys To Underscores When set to
true
, specifies that the names of the keys include dots and that you want Firehose to replace them with underscores. This is useful because Apache Hive does not allow dots in column names. For example, if the JSON contains a key whose name is "a.b", you can define the column name to be "a_b" when using this option.The default is
false
.
DeliveryStreamOrcSerDe
- Block
Size intBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- Bloom
Filter List<string>Columns - The column names for which you want Firehose to create bloom filters. The default is
null
. - Bloom
Filter doubleFalse Positive Probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- Compression string
- The compression code to use over data blocks. The default is
SNAPPY
. - Dictionary
Key doubleThreshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- Enable
Padding bool - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - Format
Version string - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - Padding
Tolerance double A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- Row
Index intStride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- Stripe
Size intBytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
- Block
Size intBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- Bloom
Filter []stringColumns - The column names for which you want Firehose to create bloom filters. The default is
null
. - Bloom
Filter float64False Positive Probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- Compression string
- The compression code to use over data blocks. The default is
SNAPPY
. - Dictionary
Key float64Threshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- Enable
Padding bool - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - Format
Version string - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - Padding
Tolerance float64 A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- Row
Index intStride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- Stripe
Size intBytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
- block
Size IntegerBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- bloom
Filter List<String>Columns - The column names for which you want Firehose to create bloom filters. The default is
null
. - bloom
Filter DoubleFalse Positive Probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- compression String
- The compression code to use over data blocks. The default is
SNAPPY
. - dictionary
Key DoubleThreshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- enable
Padding Boolean - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - format
Version String - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - padding
Tolerance Double A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- row
Index IntegerStride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- stripe
Size IntegerBytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
- block
Size numberBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- bloom
Filter string[]Columns - The column names for which you want Firehose to create bloom filters. The default is
null
. - bloom
Filter numberFalse Positive Probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- compression string
- The compression code to use over data blocks. The default is
SNAPPY
. - dictionary
Key numberThreshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- enable
Padding boolean - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - format
Version string - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - padding
Tolerance number A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- row
Index numberStride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- stripe
Size numberBytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
- block_
size_ intbytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- bloom_
filter_ Sequence[str]columns - The column names for which you want Firehose to create bloom filters. The default is
null
. - bloom_
filter_ floatfalse_ positive_ probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- compression str
- The compression code to use over data blocks. The default is
SNAPPY
. - dictionary_
key_ floatthreshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- enable_
padding bool - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - format_
version str - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - padding_
tolerance float A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- row_
index_ intstride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- stripe_
size_ intbytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
- block
Size NumberBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- bloom
Filter List<String>Columns - The column names for which you want Firehose to create bloom filters. The default is
null
. - bloom
Filter NumberFalse Positive Probability - The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.
- compression String
- The compression code to use over data blocks. The default is
SNAPPY
. - dictionary
Key NumberThreshold - Represents the fraction of the total number of non-null rows. To turn off dictionary encoding, set this fraction to a number that is less than the number of distinct keys in a dictionary. To always use dictionary encoding, set this threshold to 1.
- enable
Padding Boolean - Set this to
true
to indicate that you want stripes to be padded to the HDFS block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default isfalse
. - format
Version String - The version of the file to write. The possible values are
V0_11
andV0_12
. The default isV0_12
. - padding
Tolerance Number A number between 0 and 1 that defines the tolerance for block padding as a decimal fraction of stripe size. The default value is 0.05, which means 5 percent of stripe size.
For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB block. In such a case, if the available size within the block is more than 3.2 MiB, a new, smaller stripe is inserted to fit within that space. This ensures that no stripe crosses block boundaries and causes remote reads within a node-local task.
Kinesis Data Firehose ignores this parameter when
EnablePadding
isfalse
.- row
Index NumberStride - The number of rows between index entries. The default is 10,000 and the minimum is 1,000.
- stripe
Size NumberBytes - The number of bytes in each stripe. The default is 64 MiB and the minimum is 8 MiB.
DeliveryStreamOutputFormatConfiguration
- Serializer
Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream Serializer - Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
- Serializer
Delivery
Stream Serializer - Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
- serializer
Delivery
Stream Serializer - Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
- serializer
Delivery
Stream Serializer - Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
- serializer
Delivery
Stream Serializer - Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
- serializer Property Map
- Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet SerDe. If both are non-null, the server rejects the request.
DeliveryStreamParquetSerDe
- Block
Size intBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- Compression string
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - Enable
Dictionary boolCompression - Indicates whether to enable dictionary compression.
- Max
Padding intBytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- Page
Size intBytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- Writer
Version string - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
- Block
Size intBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- Compression string
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - Enable
Dictionary boolCompression - Indicates whether to enable dictionary compression.
- Max
Padding intBytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- Page
Size intBytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- Writer
Version string - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
- block
Size IntegerBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- compression String
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - enable
Dictionary BooleanCompression - Indicates whether to enable dictionary compression.
- max
Padding IntegerBytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- page
Size IntegerBytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- writer
Version String - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
- block
Size numberBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- compression string
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - enable
Dictionary booleanCompression - Indicates whether to enable dictionary compression.
- max
Padding numberBytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- page
Size numberBytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- writer
Version string - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
- block_
size_ intbytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- compression str
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - enable_
dictionary_ boolcompression - Indicates whether to enable dictionary compression.
- max_
padding_ intbytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- page_
size_ intbytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- writer_
version str - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
- block
Size NumberBytes - The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- compression String
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed. - enable
Dictionary BooleanCompression - Indicates whether to enable dictionary compression.
- max
Padding NumberBytes - The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- page
Size NumberBytes - The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- writer
Version String - Indicates the version of row format to output. The possible values are
V1
andV2
. The default isV1
.
DeliveryStreamProcessingConfiguration
- Enabled bool
- Indicates whether data processing is enabled (true) or disabled (false).
- Processors
List<Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processor> - The data processors.
- Enabled bool
- Indicates whether data processing is enabled (true) or disabled (false).
- Processors
[]Delivery
Stream Processor - The data processors.
- enabled Boolean
- Indicates whether data processing is enabled (true) or disabled (false).
- processors
List<Delivery
Stream Processor> - The data processors.
- enabled boolean
- Indicates whether data processing is enabled (true) or disabled (false).
- processors
Delivery
Stream Processor[] - The data processors.
- enabled bool
- Indicates whether data processing is enabled (true) or disabled (false).
- processors
Sequence[Delivery
Stream Processor] - The data processors.
- enabled Boolean
- Indicates whether data processing is enabled (true) or disabled (false).
- processors List<Property Map>
- The data processors.
DeliveryStreamProcessor
- Type
Pulumi.
Aws Native. Kinesis Firehose. Delivery Stream Processor Type - The type of processor. Valid values:
Lambda
. - Parameters
List<Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processor Parameter> - The processor parameters.
- Type
Delivery
Stream Processor Type - The type of processor. Valid values:
Lambda
. - Parameters
[]Delivery
Stream Processor Parameter - The processor parameters.
- type
Delivery
Stream Processor Type - The type of processor. Valid values:
Lambda
. - parameters
List<Delivery
Stream Processor Parameter> - The processor parameters.
- type
Delivery
Stream Processor Type - The type of processor. Valid values:
Lambda
. - parameters
Delivery
Stream Processor Parameter[] - The processor parameters.
- type
Delivery
Stream Processor Type - The type of processor. Valid values:
Lambda
. - parameters
Sequence[Delivery
Stream Processor Parameter] - The processor parameters.
- type
"Record
De Aggregation" | "Decompression" | "Cloud Watch Log Processing" | "Lambda" | "Metadata Extraction" | "Append Delimiter To Record" - The type of processor. Valid values:
Lambda
. - parameters List<Property Map>
- The processor parameters.
DeliveryStreamProcessorParameter
- Parameter
Name string - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - Parameter
Value string - The parameter value.
- Parameter
Name string - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - Parameter
Value string - The parameter value.
- parameter
Name String - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - parameter
Value String - The parameter value.
- parameter
Name string - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - parameter
Value string - The parameter value.
- parameter_
name str - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - parameter_
value str - The parameter value.
- parameter
Name String - The name of the parameter. Currently the following default values are supported: 3 for
NumberOfRetries
and 60 for theBufferIntervalInSeconds
. TheBufferSizeInMBs
ranges between 0.2 MB and up to 3MB. The default buffering hint is 1MB for all destinations, except Splunk. For Splunk, the default buffering hint is 256 KB. - parameter
Value String - The parameter value.
DeliveryStreamProcessorType
DeliveryStreamRedshiftDestinationConfiguration
- Cluster
Jdbcurl string - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- Copy
Command Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Copy Command - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - Role
Arn string - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- S3Configuration
Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream S3Destination Configuration - The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - Cloud
Watch Pulumi.Logging Options Aws Native. Kinesis Firehose. Inputs. Delivery Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- Password string
- The password for the Amazon Redshift user that you specified in the
Username
property. - Processing
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- Retry
Options Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Redshift Retry Options - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- S3Backup
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream S3Destination Configuration - The configuration for backup in Amazon S3.
- S3Backup
Mode Pulumi.Aws Native. Kinesis Firehose. Delivery Stream Redshift Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- Secrets
Manager Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Amazon Redshift.
- Username string
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
- Cluster
Jdbcurl string - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- Copy
Command DeliveryStream Copy Command - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - Role
Arn string - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- S3Configuration
Delivery
Stream S3Destination Configuration - The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - Cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- Password string
- The password for the Amazon Redshift user that you specified in the
Username
property. - Processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- Retry
Options DeliveryStream Redshift Retry Options - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- S3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- S3Backup
Mode DeliveryStream Redshift Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- Secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Amazon Redshift.
- Username string
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
- cluster
Jdbcurl String - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- copy
Command DeliveryStream Copy Command - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - role
Arn String - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- s3Configuration
Delivery
Stream S3Destination Configuration - The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- password String
- The password for the Amazon Redshift user that you specified in the
Username
property. - processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- retry
Options DeliveryStream Redshift Retry Options - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- s3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3Backup
Mode DeliveryStream Redshift Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Amazon Redshift.
- username String
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
- cluster
Jdbcurl string - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- copy
Command DeliveryStream Copy Command - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - role
Arn string - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- s3Configuration
Delivery
Stream S3Destination Configuration - The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- password string
- The password for the Amazon Redshift user that you specified in the
Username
property. - processing
Configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- retry
Options DeliveryStream Redshift Retry Options - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- s3Backup
Configuration DeliveryStream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3Backup
Mode DeliveryStream Redshift Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Amazon Redshift.
- username string
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
- cluster_
jdbcurl str - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- copy_
command DeliveryStream Copy Command - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - role_
arn str - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- s3_
configuration DeliveryStream S3Destination Configuration - The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - cloud_
watch_ Deliverylogging_ options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- password str
- The password for the Amazon Redshift user that you specified in the
Username
property. - processing_
configuration DeliveryStream Processing Configuration - The data processing configuration for the Kinesis Data Firehose delivery stream.
- retry_
options DeliveryStream Redshift Retry Options - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- s3_
backup_ Deliveryconfiguration Stream S3Destination Configuration - The configuration for backup in Amazon S3.
- s3_
backup_ Deliverymode Stream Redshift Destination Configuration S3Backup Mode - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- secrets_
manager_ Deliveryconfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Amazon Redshift.
- username str
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
- cluster
Jdbcurl String - The connection string that Kinesis Data Firehose uses to connect to the Amazon Redshift cluster.
- copy
Command Property Map - Configures the Amazon Redshift
COPY
command that Kinesis Data Firehose uses to load data into the cluster from the Amazon S3 bucket. - role
Arn String - The ARN of the AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon Redshift Destination in the Amazon Kinesis Data Firehose Developer Guide .
- s3Configuration Property Map
- The S3 bucket where Kinesis Data Firehose first delivers data. After the data is in the bucket, Kinesis Data Firehose uses the
COPY
command to load the data into the Amazon Redshift cluster. For the Amazon S3 bucket's compression format, don't specifySNAPPY
orZIP
because the Amazon RedshiftCOPY
command doesn't support them. - cloud
Watch Property MapLogging Options - The CloudWatch logging options for your Firehose stream.
- password String
- The password for the Amazon Redshift user that you specified in the
Username
property. - processing
Configuration Property Map - The data processing configuration for the Kinesis Data Firehose delivery stream.
- retry
Options Property Map - The retry behavior in case Firehose is unable to deliver documents to Amazon Redshift. Default value is 3600 (60 minutes).
- s3Backup
Configuration Property Map - The configuration for backup in Amazon S3.
- s3Backup
Mode "Disabled" | "Enabled" - The Amazon S3 backup mode. After you create a Firehose stream, you can update it to enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the Firehose stream to disable it.
- secrets
Manager Property MapConfiguration - The configuration that defines how you access secrets for Amazon Redshift.
- username String
- The Amazon Redshift user that has permission to access the Amazon Redshift cluster. This user must have
INSERT
privileges for copying data from the Amazon S3 bucket to the cluster.
DeliveryStreamRedshiftDestinationConfigurationS3BackupMode
DeliveryStreamRedshiftRetryOptions
- Duration
In intSeconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
- Duration
In intSeconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
- duration
In IntegerSeconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
- duration
In numberSeconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
- duration_
in_ intseconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
- duration
In NumberSeconds - The length of time during which Firehose retries delivery after a failure, starting from the initial request and including the first attempt. The default value is 3600 seconds (60 minutes). Firehose does not retry if the value of
DurationInSeconds
is 0 (zero) or if the first delivery attempt takes longer than the current value.
DeliveryStreamRetryOptions
- Duration
In intSeconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
- Duration
In intSeconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
- duration
In IntegerSeconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
- duration
In numberSeconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
- duration_
in_ intseconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
- duration
In NumberSeconds - The total amount of time that Kinesis Data Firehose spends on retries. This duration starts after the initial attempt to send data to the custom destination via HTTPS endpoint fails. It doesn't include the periods during which Kinesis Data Firehose waits for acknowledgment from the specified destination after each attempt.
DeliveryStreamS3DestinationConfiguration
- Bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- Role
Arn string - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- Buffering
Hints Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Buffering Hints - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- Cloud
Watch Pulumi.Logging Options Aws Native. Kinesis Firehose. Inputs. Delivery Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- Compression
Format Pulumi.Aws Native. Kinesis Firehose. Delivery Stream S3Destination Configuration Compression Format - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - Encryption
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Encryption Configuration - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- Error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- Prefix string
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
- Bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- Role
Arn string - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- Buffering
Hints DeliveryStream Buffering Hints - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- Cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- Compression
Format DeliveryStream S3Destination Configuration Compression Format - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - Encryption
Configuration DeliveryStream Encryption Configuration - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- Error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- Prefix string
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
- bucket
Arn String - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- role
Arn String - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- buffering
Hints DeliveryStream Buffering Hints - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- compression
Format DeliveryStream S3Destination Configuration Compression Format - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - encryption
Configuration DeliveryStream Encryption Configuration - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- error
Output StringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- prefix String
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
- bucket
Arn string - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- role
Arn string - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- buffering
Hints DeliveryStream Buffering Hints - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- compression
Format DeliveryStream S3Destination Configuration Compression Format - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - encryption
Configuration DeliveryStream Encryption Configuration - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- error
Output stringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- prefix string
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
- bucket_
arn str - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- role_
arn str - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- buffering_
hints DeliveryStream Buffering Hints - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- cloud_
watch_ Deliverylogging_ options Stream Cloud Watch Logging Options - The CloudWatch logging options for your Firehose stream.
- compression_
format DeliveryStream S3Destination Configuration Compression Format - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - encryption_
configuration DeliveryStream Encryption Configuration - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- error_
output_ strprefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- prefix str
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
- bucket
Arn String - The Amazon Resource Name (ARN) of the Amazon S3 bucket to send data to.
- role
Arn String - The ARN of an AWS Identity and Access Management (IAM) role that grants Kinesis Data Firehose access to your Amazon S3 bucket and AWS KMS (if you enable data encryption). For more information, see Grant Kinesis Data Firehose Access to an Amazon S3 Destination in the Amazon Kinesis Data Firehose Developer Guide .
- buffering
Hints Property Map - Configures how Kinesis Data Firehose buffers incoming data while delivering it to the Amazon S3 bucket.
- cloud
Watch Property MapLogging Options - The CloudWatch logging options for your Firehose stream.
- compression
Format "UNCOMPRESSED" | "GZIP" | "ZIP" | "Snappy" | "HADOOP_SNAPPY" - The type of compression that Kinesis Data Firehose uses to compress the data that it delivers to the Amazon S3 bucket. For valid values, see the
CompressionFormat
content for the S3DestinationConfiguration data type in the Amazon Kinesis Data Firehose API Reference . - encryption
Configuration Property Map - Configures Amazon Simple Storage Service (Amazon S3) server-side encryption. Kinesis Data Firehose uses AWS Key Management Service ( AWS KMS) to encrypt the data that it delivers to your Amazon S3 bucket.
- error
Output StringPrefix - A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing them to S3. This prefix appears immediately following the bucket name. For information about how to specify this prefix, see Custom Prefixes for Amazon S3 Objects .
- prefix String
- A prefix that Kinesis Data Firehose adds to the files that it delivers to the Amazon S3 bucket. The prefix helps you identify the files that Kinesis Data Firehose delivered.
DeliveryStreamS3DestinationConfigurationCompressionFormat
DeliveryStreamSchemaConfiguration
- Catalog
Id string - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- Database
Name string Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- Region string
- If you don't specify an AWS Region, the default is the current Region.
- Role
Arn string The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- Table
Name string Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- Version
Id string - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
- Catalog
Id string - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- Database
Name string Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- Region string
- If you don't specify an AWS Region, the default is the current Region.
- Role
Arn string The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- Table
Name string Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- Version
Id string - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
- catalog
Id String - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- database
Name String Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- region String
- If you don't specify an AWS Region, the default is the current Region.
- role
Arn String The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- table
Name String Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- version
Id String - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
- catalog
Id string - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- database
Name string Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- region string
- If you don't specify an AWS Region, the default is the current Region.
- role
Arn string The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- table
Name string Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- version
Id string - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
- catalog_
id str - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- database_
name str Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- region str
- If you don't specify an AWS Region, the default is the current Region.
- role_
arn str The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- table_
name str Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- version_
id str - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
- catalog
Id String - The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is used by default.
- database
Name String Specifies the name of the AWS Glue database that contains the schema for the output data.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theDatabaseName
property is required and its value must be specified.- region String
- If you don't specify an AWS Region, the default is the current Region.
- role
Arn String The role that Firehose can use to access AWS Glue. This role must be in the same account you use for Firehose. Cross-account roles aren't allowed.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theRoleARN
property is required and its value must be specified.- table
Name String Specifies the AWS Glue table that contains the column information that constitutes your data schema.
If the
SchemaConfiguration
request parameter is used as part of invoking theCreateDeliveryStream
API, then theTableName
property is required and its value must be specified.- version
Id String - Specifies the table version for the output data schema. If you don't specify this version ID, or if you set it to
LATEST
, Firehose uses the most recent version. This means that any updates to the table are automatically picked up.
DeliveryStreamSecretsManagerConfiguration
- Enabled bool
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - Role
Arn string - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- Secret
Arn string - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
- Enabled bool
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - Role
Arn string - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- Secret
Arn string - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
- enabled Boolean
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - role
Arn String - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- secret
Arn String - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
- enabled boolean
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - role
Arn string - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- secret
Arn string - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
- enabled bool
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - role_
arn str - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- secret_
arn str - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
- enabled Boolean
- Specifies whether you want to use the secrets manager feature. When set as
True
the secrets manager configuration overwrites the existing secrets in the destination configuration. When it's set toFalse
Firehose falls back to the credentials in the destination configuration. - role
Arn String - Specifies the role that Firehose assumes when calling the Secrets Manager API operation. When you provide the role, it overrides any destination specific role defined in the destination configuration. If you do not provide the then we use the destination specific role. This parameter is required for Splunk.
- secret
Arn String - The ARN of the secret that stores your credentials. It must be in the same region as the Firehose stream and the role. The secret ARN can reside in a different account than the Firehose stream and role as Firehose supports cross-account secret access. This parameter is required when Enabled is set to
True
.
DeliveryStreamSerializer
- Orc
Ser Pulumi.De Aws Native. Kinesis Firehose. Inputs. Delivery Stream Orc Ser De - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- Parquet
Ser Pulumi.De Aws Native. Kinesis Firehose. Inputs. Delivery Stream Parquet Ser De - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
- Orc
Ser DeliveryDe Stream Orc Ser De - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- Parquet
Ser DeliveryDe Stream Parquet Ser De - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
- orc
Ser DeliveryDe Stream Orc Ser De - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- parquet
Ser DeliveryDe Stream Parquet Ser De - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
- orc
Ser DeliveryDe Stream Orc Ser De - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- parquet
Ser DeliveryDe Stream Parquet Ser De - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
- orc_
ser_ Deliveryde Stream Orc Ser De - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- parquet_
ser_ Deliveryde Stream Parquet Ser De - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
- orc
Ser Property MapDe - A serializer to use for converting data to the ORC format before storing it in Amazon S3. For more information, see Apache ORC .
- parquet
Ser Property MapDe - A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet .
DeliveryStreamSplunkBufferingHints
- Interval
In intSeconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- Size
In intMbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
- Interval
In intSeconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- Size
In intMbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
- interval
In IntegerSeconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- size
In IntegerMbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
- interval
In numberSeconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- size
In numberMbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
- interval_
in_ intseconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- size_
in_ intmbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
- interval
In NumberSeconds - Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).
- size
In NumberMbs - Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5.
DeliveryStreamSplunkDestinationConfiguration
- Hec
Endpoint string - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- Hec
Endpoint Pulumi.Type Aws Native. Kinesis Firehose. Delivery Stream Splunk Destination Configuration Hec Endpoint Type - This type can be either
Raw
orEvent
. - S3Configuration
Pulumi.
Aws Native. Kinesis Firehose. Inputs. Delivery Stream S3Destination Configuration - The configuration for the backup Amazon S3 location.
- Buffering
Hints Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Splunk Buffering Hints - The buffering options. If no value is specified, the default values for Splunk are used.
- Cloud
Watch Pulumi.Logging Options Aws Native. Kinesis Firehose. Inputs. Delivery Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- Hec
Acknowledgment intTimeout In Seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- Hec
Token string - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- Processing
Configuration Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Processing Configuration - The data processing configuration.
- Retry
Options Pulumi.Aws Native. Kinesis Firehose. Inputs. Delivery Stream Splunk Retry Options - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- S3Backup
Mode string Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- Secrets
Manager Pulumi.Configuration Aws Native. Kinesis Firehose. Inputs. Delivery Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Splunk.
- Hec
Endpoint string - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- Hec
Endpoint DeliveryType Stream Splunk Destination Configuration Hec Endpoint Type - This type can be either
Raw
orEvent
. - S3Configuration
Delivery
Stream S3Destination Configuration - The configuration for the backup Amazon S3 location.
- Buffering
Hints DeliveryStream Splunk Buffering Hints - The buffering options. If no value is specified, the default values for Splunk are used.
- Cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- Hec
Acknowledgment intTimeout In Seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- Hec
Token string - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- Processing
Configuration DeliveryStream Processing Configuration - The data processing configuration.
- Retry
Options DeliveryStream Splunk Retry Options - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- S3Backup
Mode string Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- Secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Splunk.
- hec
Endpoint String - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- hec
Endpoint DeliveryType Stream Splunk Destination Configuration Hec Endpoint Type - This type can be either
Raw
orEvent
. - s3Configuration
Delivery
Stream S3Destination Configuration - The configuration for the backup Amazon S3 location.
- buffering
Hints DeliveryStream Splunk Buffering Hints - The buffering options. If no value is specified, the default values for Splunk are used.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- hec
Acknowledgment IntegerTimeout In Seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- hec
Token String - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- processing
Configuration DeliveryStream Processing Configuration - The data processing configuration.
- retry
Options DeliveryStream Splunk Retry Options - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- s3Backup
Mode String Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Splunk.
- hec
Endpoint string - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- hec
Endpoint DeliveryType Stream Splunk Destination Configuration Hec Endpoint Type - This type can be either
Raw
orEvent
. - s3Configuration
Delivery
Stream S3Destination Configuration - The configuration for the backup Amazon S3 location.
- buffering
Hints DeliveryStream Splunk Buffering Hints - The buffering options. If no value is specified, the default values for Splunk are used.
- cloud
Watch DeliveryLogging Options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- hec
Acknowledgment numberTimeout In Seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- hec
Token string - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- processing
Configuration DeliveryStream Processing Configuration - The data processing configuration.
- retry
Options DeliveryStream Splunk Retry Options - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- s3Backup
Mode string Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- secrets
Manager DeliveryConfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Splunk.
- hec_
endpoint str - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- hec_
endpoint_ Deliverytype Stream Splunk Destination Configuration Hec Endpoint Type - This type can be either
Raw
orEvent
. - s3_
configuration DeliveryStream S3Destination Configuration - The configuration for the backup Amazon S3 location.
- buffering_
hints DeliveryStream Splunk Buffering Hints - The buffering options. If no value is specified, the default values for Splunk are used.
- cloud_
watch_ Deliverylogging_ options Stream Cloud Watch Logging Options - The Amazon CloudWatch logging options for your Firehose stream.
- hec_
acknowledgment_ inttimeout_ in_ seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- hec_
token str - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- processing_
configuration DeliveryStream Processing Configuration - The data processing configuration.
- retry_
options DeliveryStream Splunk Retry Options - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- s3_
backup_ strmode Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- secrets_
manager_ Deliveryconfiguration Stream Secrets Manager Configuration - The configuration that defines how you access secrets for Splunk.
- hec
Endpoint String - The HTTP Event Collector (HEC) endpoint to which Firehose sends your data.
- hec
Endpoint "Raw" | "Event"Type - This type can be either
Raw
orEvent
. - s3Configuration Property Map
- The configuration for the backup Amazon S3 location.
- buffering
Hints Property Map - The buffering options. If no value is specified, the default values for Splunk are used.
- cloud
Watch Property MapLogging Options - The Amazon CloudWatch logging options for your Firehose stream.
- hec
Acknowledgment NumberTimeout In Seconds - The amount of time that Firehose waits to receive an acknowledgment from Splunk after it sends it data. At the end of the timeout period, Firehose either tries to send the data again or considers it an error, based on your retry settings.
- hec
Token String - This is a GUID that you obtain from your Splunk cluster when you create a new HEC endpoint.
- processing
Configuration Property Map - The data processing configuration.
- retry
Options Property Map - The retry behavior in case Firehose is unable to deliver data to Splunk, or if it doesn't receive an acknowledgment of receipt from Splunk.
- s3Backup
Mode String Defines how documents should be delivered to Amazon S3. When set to
FailedEventsOnly
, Firehose writes any data that could not be indexed to the configured Amazon S3 destination. When set toAllEvents
, Firehose delivers all incoming records to Amazon S3, and also writes failed documents to Amazon S3. The default value isFailedEventsOnly
.You can update this backup mode from
FailedEventsOnly
toAllEvents
. You can't update it fromAllEvents
toFailedEventsOnly
.- secrets
Manager Property MapConfiguration - The configuration that defines how you access secrets for Splunk.
DeliveryStreamSplunkDestinationConfigurationHecEndpointType
DeliveryStreamSplunkRetryOptions
- Duration
In intSeconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
- Duration
In intSeconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
- duration
In IntegerSeconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
- duration
In numberSeconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
- duration_
in_ intseconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
- duration
In NumberSeconds - The total amount of time that Firehose spends on retries. This duration starts after the initial attempt to send data to Splunk fails. It doesn't include the periods during which Firehose waits for acknowledgment from Splunk after each attempt.
Tag
Package Details
- Repository
- AWS Native pulumi/pulumi-aws-native
- License
- Apache-2.0
We recommend new projects start with resources from the AWS provider.