diff --git a/.generator/schemas/v2/openapi.yaml b/.generator/schemas/v2/openapi.yaml index 81b1f9b3a2..6a18584144 100644 --- a/.generator/schemas/v2/openapi.yaml +++ b/.generator/schemas/v2/openapi.yaml @@ -41994,6 +41994,191 @@ components: type: string x-enum-varnames: - AMAZON_S3 + ObservabilityPipelineAmazonS3GenericBatchSettings: + description: Event batching settings + properties: + batch_size: + description: Maximum batch size in bytes. + example: 100000000 + format: int64 + type: integer + timeout_secs: + description: Maximum number of seconds to wait before flushing the batch. + example: 900 + format: int64 + type: integer + type: object + ObservabilityPipelineAmazonS3GenericCompression: + description: Compression algorithm applied to encoded logs. + oneOf: + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionZstd' + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionGzip' + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionSnappy' + ObservabilityPipelineAmazonS3GenericCompressionGzip: + description: Gzip compression. + properties: + level: + description: Gzip compression level. + example: 6 + format: int64 + type: integer + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionGzipType' + required: + - type + - level + type: object + ObservabilityPipelineAmazonS3GenericCompressionGzipType: + default: gzip + description: The compression type. Always `gzip`. + enum: + - gzip + example: gzip + type: string + x-enum-varnames: + - GZIP + ObservabilityPipelineAmazonS3GenericCompressionSnappy: + description: Snappy compression. + properties: + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionSnappyType' + required: + - type + type: object + ObservabilityPipelineAmazonS3GenericCompressionSnappyType: + default: snappy + description: The compression type. Always `snappy`. + enum: + - snappy + example: snappy + type: string + x-enum-varnames: + - SNAPPY + ObservabilityPipelineAmazonS3GenericCompressionZstd: + description: Zstd compression. + properties: + level: + description: Zstd compression level. + example: 3 + format: int64 + type: integer + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompressionZstdType' + required: + - type + - level + type: object + ObservabilityPipelineAmazonS3GenericCompressionZstdType: + default: zstd + description: The compression type. Always `zstd`. + enum: + - zstd + example: zstd + type: string + x-enum-varnames: + - ZSTD + ObservabilityPipelineAmazonS3GenericDestination: + description: 'The `amazon_s3_generic` destination sends your logs to an Amazon + S3 bucket. + + + **Supported pipeline types:** logs' + properties: + auth: + $ref: '#/components/schemas/ObservabilityPipelineAwsAuth' + batch_settings: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericBatchSettings' + bucket: + description: S3 bucket name. + example: my-bucket + type: string + compression: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericCompression' + encoding: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericEncoding' + id: + description: Unique identifier for the destination component. + example: generic-s3-destination + type: string + inputs: + description: A list of component IDs whose output is used as the `input` + for this component. + example: + - filter-processor + items: + type: string + type: array + key_prefix: + description: Optional prefix for object keys. + type: string + region: + description: AWS region of the S3 bucket. + example: us-east-1 + type: string + storage_class: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3DestinationStorageClass' + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericDestinationType' + required: + - id + - type + - inputs + - bucket + - region + - storage_class + - encoding + - compression + type: object + x-pipeline-types: + - logs + ObservabilityPipelineAmazonS3GenericDestinationType: + default: amazon_s3_generic + description: The destination type. Always `amazon_s3_generic`. + enum: + - amazon_s3_generic + example: amazon_s3_generic + type: string + x-enum-varnames: + - GENERIC_ARCHIVES_S3 + ObservabilityPipelineAmazonS3GenericEncoding: + description: Encoding format for the destination. + oneOf: + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericEncodingJson' + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericEncodingParquet' + ObservabilityPipelineAmazonS3GenericEncodingJson: + description: JSON encoding. + properties: + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericEncodingJsonType' + required: + - type + type: object + ObservabilityPipelineAmazonS3GenericEncodingJsonType: + default: json + description: The encoding type. Always `json`. + enum: + - json + example: json + type: string + x-enum-varnames: + - JSON + ObservabilityPipelineAmazonS3GenericEncodingParquet: + description: Parquet encoding. + properties: + type: + $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericEncodingParquetType' + required: + - type + type: object + ObservabilityPipelineAmazonS3GenericEncodingParquetType: + default: parquet + description: The encoding type. Always `parquet`. + enum: + - parquet + example: parquet + type: string + x-enum-varnames: + - PARQUET ObservabilityPipelineAmazonS3Source: description: 'The `amazon_s3` source ingests logs from an Amazon S3 bucket. @@ -42269,6 +42454,7 @@ components: - $ref: '#/components/schemas/ObservabilityPipelineHttpClientDestination' - $ref: '#/components/schemas/ObservabilityPipelineAmazonOpenSearchDestination' - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3Destination' + - $ref: '#/components/schemas/ObservabilityPipelineAmazonS3GenericDestination' - $ref: '#/components/schemas/ObservabilityPipelineAmazonSecurityLakeDestination' - $ref: '#/components/schemas/AzureStorageDestination' - $ref: '#/components/schemas/ObservabilityPipelineCloudPremDestination' @@ -46462,15 +46648,6 @@ components: description: Optional name of the Splunk index where logs are written. example: main type: string - indexed_fields: - description: List of log field names to send as indexed fields to Splunk - HEC. Available only when `encoding` is `json`. - example: - - service - - host - items: - type: string - type: array inputs: description: A list of component IDs whose output is used as the `input` for this component. diff --git a/docs/datadog_api_client.v2.model.rst b/docs/datadog_api_client.v2.model.rst index c312689ac6..88fd535ef5 100644 --- a/docs/datadog_api_client.v2.model.rst +++ b/docs/datadog_api_client.v2.model.rst @@ -18365,6 +18365,111 @@ datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_destination\_ :members: :show-inheritance: +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_batch\_settings module +-------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_batch_settings + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression module +---------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_gzip module +---------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_gzip\_type module +---------------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip_type + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_snappy module +------------------------------------------------------------------------------------------------------ + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_snappy\_type module +------------------------------------------------------------------------------------------------------------ + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy_type + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_zstd module +---------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_compression\_zstd\_type module +---------------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd_type + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_destination module +---------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_destination\_type module +---------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination_type + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_encoding module +------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_encoding\_json module +------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_encoding\_json\_type module +------------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json_type + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_encoding\_parquet module +---------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet + :members: + :show-inheritance: + +datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_generic\_encoding\_parquet\_type module +---------------------------------------------------------------------------------------------------------- + +.. automodule:: datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet_type + :members: + :show-inheritance: + datadog\_api\_client.v2.model.observability\_pipeline\_amazon\_s3\_source module -------------------------------------------------------------------------------- diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_batch_settings.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_batch_settings.py new file mode 100644 index 0000000000..3a3578446f --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_batch_settings.py @@ -0,0 +1,45 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import Union + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, + unset, + UnsetType, +) + + +class ObservabilityPipelineAmazonS3GenericBatchSettings(ModelNormal): + @cached_property + def openapi_types(_): + return { + "batch_size": (int,), + "timeout_secs": (int,), + } + + attribute_map = { + "batch_size": "batch_size", + "timeout_secs": "timeout_secs", + } + + def __init__( + self_, batch_size: Union[int, UnsetType] = unset, timeout_secs: Union[int, UnsetType] = unset, **kwargs + ): + """ + Event batching settings + + :param batch_size: Maximum batch size in bytes. + :type batch_size: int, optional + + :param timeout_secs: Maximum number of seconds to wait before flushing the batch. + :type timeout_secs: int, optional + """ + if batch_size is not unset: + kwargs["batch_size"] = batch_size + if timeout_secs is not unset: + kwargs["timeout_secs"] = timeout_secs + super().__init__(kwargs) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression.py new file mode 100644 index 0000000000..81f726cb03 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression.py @@ -0,0 +1,51 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelComposed, + cached_property, +) + + +class ObservabilityPipelineAmazonS3GenericCompression(ModelComposed): + def __init__(self, **kwargs): + """ + Compression algorithm applied to encoded logs. + + :param level: Zstd compression level. + :type level: int + + :param type: The compression type. Always `zstd`. + :type type: ObservabilityPipelineAmazonS3GenericCompressionZstdType + """ + super().__init__(kwargs) + + @cached_property + def _composed_schemas(_): + # we need this here to make our import statements work + # we must store _composed_schemas in here so the code is only run + # when we invoke this method. If we kept this at the class + # level we would get an error because the class level + # code would be run when this module is imported, and these composed + # classes don't exist yet because their module has not finished + # loading + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd import ( + ObservabilityPipelineAmazonS3GenericCompressionZstd, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip import ( + ObservabilityPipelineAmazonS3GenericCompressionGzip, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappy, + ) + + return { + "oneOf": [ + ObservabilityPipelineAmazonS3GenericCompressionZstd, + ObservabilityPipelineAmazonS3GenericCompressionGzip, + ObservabilityPipelineAmazonS3GenericCompressionSnappy, + ], + } diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip.py new file mode 100644 index 0000000000..97c120d26b --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip.py @@ -0,0 +1,50 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip_type import ( + ObservabilityPipelineAmazonS3GenericCompressionGzipType, + ) + + +class ObservabilityPipelineAmazonS3GenericCompressionGzip(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip_type import ( + ObservabilityPipelineAmazonS3GenericCompressionGzipType, + ) + + return { + "level": (int,), + "type": (ObservabilityPipelineAmazonS3GenericCompressionGzipType,), + } + + attribute_map = { + "level": "level", + "type": "type", + } + + def __init__(self_, level: int, type: ObservabilityPipelineAmazonS3GenericCompressionGzipType, **kwargs): + """ + Gzip compression. + + :param level: Gzip compression level. + :type level: int + + :param type: The compression type. Always ``gzip``. + :type type: ObservabilityPipelineAmazonS3GenericCompressionGzipType + """ + super().__init__(kwargs) + + self_.level = level + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip_type.py new file mode 100644 index 0000000000..3f8d2b4853 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_gzip_type.py @@ -0,0 +1,37 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericCompressionGzipType(ModelSimple): + """ + The compression type. Always `gzip`. + + :param value: If omitted defaults to "gzip". Must be one of ["gzip"]. + :type value: str + """ + + allowed_values = { + "gzip", + } + GZIP: ClassVar["ObservabilityPipelineAmazonS3GenericCompressionGzipType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericCompressionGzipType.GZIP = ObservabilityPipelineAmazonS3GenericCompressionGzipType( + "gzip" +) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy.py new file mode 100644 index 0000000000..42d14349bc --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy.py @@ -0,0 +1,44 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy_type import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappyType, + ) + + +class ObservabilityPipelineAmazonS3GenericCompressionSnappy(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy_type import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappyType, + ) + + return { + "type": (ObservabilityPipelineAmazonS3GenericCompressionSnappyType,), + } + + attribute_map = { + "type": "type", + } + + def __init__(self_, type: ObservabilityPipelineAmazonS3GenericCompressionSnappyType, **kwargs): + """ + Snappy compression. + + :param type: The compression type. Always ``snappy``. + :type type: ObservabilityPipelineAmazonS3GenericCompressionSnappyType + """ + super().__init__(kwargs) + + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy_type.py new file mode 100644 index 0000000000..107ed8da5d --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_snappy_type.py @@ -0,0 +1,37 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericCompressionSnappyType(ModelSimple): + """ + The compression type. Always `snappy`. + + :param value: If omitted defaults to "snappy". Must be one of ["snappy"]. + :type value: str + """ + + allowed_values = { + "snappy", + } + SNAPPY: ClassVar["ObservabilityPipelineAmazonS3GenericCompressionSnappyType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericCompressionSnappyType.SNAPPY = ( + ObservabilityPipelineAmazonS3GenericCompressionSnappyType("snappy") +) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd.py new file mode 100644 index 0000000000..af449a140f --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd.py @@ -0,0 +1,50 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd_type import ( + ObservabilityPipelineAmazonS3GenericCompressionZstdType, + ) + + +class ObservabilityPipelineAmazonS3GenericCompressionZstd(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd_type import ( + ObservabilityPipelineAmazonS3GenericCompressionZstdType, + ) + + return { + "level": (int,), + "type": (ObservabilityPipelineAmazonS3GenericCompressionZstdType,), + } + + attribute_map = { + "level": "level", + "type": "type", + } + + def __init__(self_, level: int, type: ObservabilityPipelineAmazonS3GenericCompressionZstdType, **kwargs): + """ + Zstd compression. + + :param level: Zstd compression level. + :type level: int + + :param type: The compression type. Always ``zstd``. + :type type: ObservabilityPipelineAmazonS3GenericCompressionZstdType + """ + super().__init__(kwargs) + + self_.level = level + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd_type.py new file mode 100644 index 0000000000..ffea9d632c --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_compression_zstd_type.py @@ -0,0 +1,37 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericCompressionZstdType(ModelSimple): + """ + The compression type. Always `zstd`. + + :param value: If omitted defaults to "zstd". Must be one of ["zstd"]. + :type value: str + """ + + allowed_values = { + "zstd", + } + ZSTD: ClassVar["ObservabilityPipelineAmazonS3GenericCompressionZstdType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericCompressionZstdType.ZSTD = ObservabilityPipelineAmazonS3GenericCompressionZstdType( + "zstd" +) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination.py new file mode 100644 index 0000000000..2c7e740af7 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination.py @@ -0,0 +1,176 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import List, Union, TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, + unset, + UnsetType, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_aws_auth import ObservabilityPipelineAwsAuth + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_batch_settings import ( + ObservabilityPipelineAmazonS3GenericBatchSettings, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression import ( + ObservabilityPipelineAmazonS3GenericCompression, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding import ( + ObservabilityPipelineAmazonS3GenericEncoding, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_destination_storage_class import ( + ObservabilityPipelineAmazonS3DestinationStorageClass, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination_type import ( + ObservabilityPipelineAmazonS3GenericDestinationType, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd import ( + ObservabilityPipelineAmazonS3GenericCompressionZstd, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip import ( + ObservabilityPipelineAmazonS3GenericCompressionGzip, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappy, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json import ( + ObservabilityPipelineAmazonS3GenericEncodingJson, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet import ( + ObservabilityPipelineAmazonS3GenericEncodingParquet, + ) + + +class ObservabilityPipelineAmazonS3GenericDestination(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_aws_auth import ObservabilityPipelineAwsAuth + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_batch_settings import ( + ObservabilityPipelineAmazonS3GenericBatchSettings, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression import ( + ObservabilityPipelineAmazonS3GenericCompression, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding import ( + ObservabilityPipelineAmazonS3GenericEncoding, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_destination_storage_class import ( + ObservabilityPipelineAmazonS3DestinationStorageClass, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination_type import ( + ObservabilityPipelineAmazonS3GenericDestinationType, + ) + + return { + "auth": (ObservabilityPipelineAwsAuth,), + "batch_settings": (ObservabilityPipelineAmazonS3GenericBatchSettings,), + "bucket": (str,), + "compression": (ObservabilityPipelineAmazonS3GenericCompression,), + "encoding": (ObservabilityPipelineAmazonS3GenericEncoding,), + "id": (str,), + "inputs": ([str],), + "key_prefix": (str,), + "region": (str,), + "storage_class": (ObservabilityPipelineAmazonS3DestinationStorageClass,), + "type": (ObservabilityPipelineAmazonS3GenericDestinationType,), + } + + attribute_map = { + "auth": "auth", + "batch_settings": "batch_settings", + "bucket": "bucket", + "compression": "compression", + "encoding": "encoding", + "id": "id", + "inputs": "inputs", + "key_prefix": "key_prefix", + "region": "region", + "storage_class": "storage_class", + "type": "type", + } + + def __init__( + self_, + bucket: str, + compression: Union[ + ObservabilityPipelineAmazonS3GenericCompression, + ObservabilityPipelineAmazonS3GenericCompressionZstd, + ObservabilityPipelineAmazonS3GenericCompressionGzip, + ObservabilityPipelineAmazonS3GenericCompressionSnappy, + ], + encoding: Union[ + ObservabilityPipelineAmazonS3GenericEncoding, + ObservabilityPipelineAmazonS3GenericEncodingJson, + ObservabilityPipelineAmazonS3GenericEncodingParquet, + ], + id: str, + inputs: List[str], + region: str, + storage_class: ObservabilityPipelineAmazonS3DestinationStorageClass, + type: ObservabilityPipelineAmazonS3GenericDestinationType, + auth: Union[ObservabilityPipelineAwsAuth, UnsetType] = unset, + batch_settings: Union[ObservabilityPipelineAmazonS3GenericBatchSettings, UnsetType] = unset, + key_prefix: Union[str, UnsetType] = unset, + **kwargs, + ): + """ + The ``amazon_s3_generic`` destination sends your logs to an Amazon S3 bucket. + + **Supported pipeline types:** logs + + :param auth: AWS authentication credentials used for accessing AWS services such as S3. + If omitted, the system’s default credentials are used (for example, the IAM role and environment variables). + :type auth: ObservabilityPipelineAwsAuth, optional + + :param batch_settings: Event batching settings + :type batch_settings: ObservabilityPipelineAmazonS3GenericBatchSettings, optional + + :param bucket: S3 bucket name. + :type bucket: str + + :param compression: Compression algorithm applied to encoded logs. + :type compression: ObservabilityPipelineAmazonS3GenericCompression + + :param encoding: Encoding format for the destination. + :type encoding: ObservabilityPipelineAmazonS3GenericEncoding + + :param id: Unique identifier for the destination component. + :type id: str + + :param inputs: A list of component IDs whose output is used as the ``input`` for this component. + :type inputs: [str] + + :param key_prefix: Optional prefix for object keys. + :type key_prefix: str, optional + + :param region: AWS region of the S3 bucket. + :type region: str + + :param storage_class: S3 storage class. + :type storage_class: ObservabilityPipelineAmazonS3DestinationStorageClass + + :param type: The destination type. Always ``amazon_s3_generic``. + :type type: ObservabilityPipelineAmazonS3GenericDestinationType + """ + if auth is not unset: + kwargs["auth"] = auth + if batch_settings is not unset: + kwargs["batch_settings"] = batch_settings + if key_prefix is not unset: + kwargs["key_prefix"] = key_prefix + super().__init__(kwargs) + + self_.bucket = bucket + self_.compression = compression + self_.encoding = encoding + self_.id = id + self_.inputs = inputs + self_.region = region + self_.storage_class = storage_class + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination_type.py new file mode 100644 index 0000000000..79bd32ac6f --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_destination_type.py @@ -0,0 +1,37 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericDestinationType(ModelSimple): + """ + The destination type. Always `amazon_s3_generic`. + + :param value: If omitted defaults to "amazon_s3_generic". Must be one of ["amazon_s3_generic"]. + :type value: str + """ + + allowed_values = { + "amazon_s3_generic", + } + GENERIC_ARCHIVES_S3: ClassVar["ObservabilityPipelineAmazonS3GenericDestinationType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericDestinationType.GENERIC_ARCHIVES_S3 = ( + ObservabilityPipelineAmazonS3GenericDestinationType("amazon_s3_generic") +) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding.py new file mode 100644 index 0000000000..2a2f08cf44 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding.py @@ -0,0 +1,44 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelComposed, + cached_property, +) + + +class ObservabilityPipelineAmazonS3GenericEncoding(ModelComposed): + def __init__(self, **kwargs): + """ + Encoding format for the destination. + + :param type: The encoding type. Always `json`. + :type type: ObservabilityPipelineAmazonS3GenericEncodingJsonType + """ + super().__init__(kwargs) + + @cached_property + def _composed_schemas(_): + # we need this here to make our import statements work + # we must store _composed_schemas in here so the code is only run + # when we invoke this method. If we kept this at the class + # level we would get an error because the class level + # code would be run when this module is imported, and these composed + # classes don't exist yet because their module has not finished + # loading + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json import ( + ObservabilityPipelineAmazonS3GenericEncodingJson, + ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet import ( + ObservabilityPipelineAmazonS3GenericEncodingParquet, + ) + + return { + "oneOf": [ + ObservabilityPipelineAmazonS3GenericEncodingJson, + ObservabilityPipelineAmazonS3GenericEncodingParquet, + ], + } diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json.py new file mode 100644 index 0000000000..c02ebf1543 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json.py @@ -0,0 +1,44 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json_type import ( + ObservabilityPipelineAmazonS3GenericEncodingJsonType, + ) + + +class ObservabilityPipelineAmazonS3GenericEncodingJson(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json_type import ( + ObservabilityPipelineAmazonS3GenericEncodingJsonType, + ) + + return { + "type": (ObservabilityPipelineAmazonS3GenericEncodingJsonType,), + } + + attribute_map = { + "type": "type", + } + + def __init__(self_, type: ObservabilityPipelineAmazonS3GenericEncodingJsonType, **kwargs): + """ + JSON encoding. + + :param type: The encoding type. Always ``json``. + :type type: ObservabilityPipelineAmazonS3GenericEncodingJsonType + """ + super().__init__(kwargs) + + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json_type.py new file mode 100644 index 0000000000..adc6406c78 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_json_type.py @@ -0,0 +1,35 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericEncodingJsonType(ModelSimple): + """ + The encoding type. Always `json`. + + :param value: If omitted defaults to "json". Must be one of ["json"]. + :type value: str + """ + + allowed_values = { + "json", + } + JSON: ClassVar["ObservabilityPipelineAmazonS3GenericEncodingJsonType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericEncodingJsonType.JSON = ObservabilityPipelineAmazonS3GenericEncodingJsonType("json") diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet.py new file mode 100644 index 0000000000..8ce01f6fb0 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet.py @@ -0,0 +1,44 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from datadog_api_client.model_utils import ( + ModelNormal, + cached_property, +) + + +if TYPE_CHECKING: + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet_type import ( + ObservabilityPipelineAmazonS3GenericEncodingParquetType, + ) + + +class ObservabilityPipelineAmazonS3GenericEncodingParquet(ModelNormal): + @cached_property + def openapi_types(_): + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet_type import ( + ObservabilityPipelineAmazonS3GenericEncodingParquetType, + ) + + return { + "type": (ObservabilityPipelineAmazonS3GenericEncodingParquetType,), + } + + attribute_map = { + "type": "type", + } + + def __init__(self_, type: ObservabilityPipelineAmazonS3GenericEncodingParquetType, **kwargs): + """ + Parquet encoding. + + :param type: The encoding type. Always ``parquet``. + :type type: ObservabilityPipelineAmazonS3GenericEncodingParquetType + """ + super().__init__(kwargs) + + self_.type = type diff --git a/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet_type.py b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet_type.py new file mode 100644 index 0000000000..2e79b0bfd7 --- /dev/null +++ b/src/datadog_api_client/v2/model/observability_pipeline_amazon_s3_generic_encoding_parquet_type.py @@ -0,0 +1,37 @@ +# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. +# This product includes software developed at Datadog (https://www.datadoghq.com/). +# Copyright 2019-Present Datadog, Inc. +from __future__ import annotations + + +from datadog_api_client.model_utils import ( + ModelSimple, + cached_property, +) + +from typing import ClassVar + + +class ObservabilityPipelineAmazonS3GenericEncodingParquetType(ModelSimple): + """ + The encoding type. Always `parquet`. + + :param value: If omitted defaults to "parquet". Must be one of ["parquet"]. + :type value: str + """ + + allowed_values = { + "parquet", + } + PARQUET: ClassVar["ObservabilityPipelineAmazonS3GenericEncodingParquetType"] + + @cached_property + def openapi_types(_): + return { + "value": (str,), + } + + +ObservabilityPipelineAmazonS3GenericEncodingParquetType.PARQUET = ( + ObservabilityPipelineAmazonS3GenericEncodingParquetType("parquet") +) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_config.py b/src/datadog_api_client/v2/model/observability_pipeline_config.py index bc983ec82a..e5b0694896 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_config.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_config.py @@ -35,6 +35,9 @@ from datadog_api_client.v2.model.observability_pipeline_amazon_s3_destination import ( ObservabilityPipelineAmazonS3Destination, ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination import ( + ObservabilityPipelineAmazonS3GenericDestination, + ) from datadog_api_client.v2.model.observability_pipeline_amazon_security_lake_destination import ( ObservabilityPipelineAmazonSecurityLakeDestination, ) @@ -172,6 +175,7 @@ def __init__( ObservabilityPipelineHttpClientDestination, ObservabilityPipelineAmazonOpenSearchDestination, ObservabilityPipelineAmazonS3Destination, + ObservabilityPipelineAmazonS3GenericDestination, ObservabilityPipelineAmazonSecurityLakeDestination, AzureStorageDestination, ObservabilityPipelineCloudPremDestination, diff --git a/src/datadog_api_client/v2/model/observability_pipeline_config_destination_item.py b/src/datadog_api_client/v2/model/observability_pipeline_config_destination_item.py index 1fe985ad30..87eb98b445 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_config_destination_item.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_config_destination_item.py @@ -73,6 +73,9 @@ def __init__(self, **kwargs): :param storage_class: S3 storage class. :type storage_class: ObservabilityPipelineAmazonS3DestinationStorageClass + :param batch_settings: Event batching settings + :type batch_settings: ObservabilityPipelineAmazonS3GenericBatchSettings, optional + :param custom_source_name: Custom source name for the logs in Security Lake. :type custom_source_name: str @@ -185,9 +188,6 @@ def __init__(self, **kwargs): :param index: Optional name of the Splunk index where logs are written. :type index: str, optional - :param indexed_fields: List of log field names to send as indexed fields to Splunk HEC. Available only when `encoding` is `json`. - :type indexed_fields: [str], optional - :param sourcetype: The Splunk sourcetype to assign to log events. :type sourcetype: str, optional @@ -223,6 +223,9 @@ def _composed_schemas(_): from datadog_api_client.v2.model.observability_pipeline_amazon_s3_destination import ( ObservabilityPipelineAmazonS3Destination, ) + from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination import ( + ObservabilityPipelineAmazonS3GenericDestination, + ) from datadog_api_client.v2.model.observability_pipeline_amazon_security_lake_destination import ( ObservabilityPipelineAmazonSecurityLakeDestination, ) @@ -285,6 +288,7 @@ def _composed_schemas(_): ObservabilityPipelineHttpClientDestination, ObservabilityPipelineAmazonOpenSearchDestination, ObservabilityPipelineAmazonS3Destination, + ObservabilityPipelineAmazonS3GenericDestination, ObservabilityPipelineAmazonSecurityLakeDestination, AzureStorageDestination, ObservabilityPipelineCloudPremDestination, diff --git a/src/datadog_api_client/v2/model/observability_pipeline_splunk_hec_destination.py b/src/datadog_api_client/v2/model/observability_pipeline_splunk_hec_destination.py index bde81cfef4..ab381977ff 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_splunk_hec_destination.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_splunk_hec_destination.py @@ -50,7 +50,6 @@ def openapi_types(_): "endpoint_url_key": (str,), "id": (str,), "index": (str,), - "indexed_fields": ([str],), "inputs": ([str],), "sourcetype": (str,), "token_key": (str,), @@ -64,7 +63,6 @@ def openapi_types(_): "endpoint_url_key": "endpoint_url_key", "id": "id", "index": "index", - "indexed_fields": "indexed_fields", "inputs": "inputs", "sourcetype": "sourcetype", "token_key": "token_key", @@ -87,7 +85,6 @@ def __init__( encoding: Union[ObservabilityPipelineSplunkHecDestinationEncoding, UnsetType] = unset, endpoint_url_key: Union[str, UnsetType] = unset, index: Union[str, UnsetType] = unset, - indexed_fields: Union[List[str], UnsetType] = unset, sourcetype: Union[str, UnsetType] = unset, token_key: Union[str, UnsetType] = unset, **kwargs, @@ -116,9 +113,6 @@ def __init__( :param index: Optional name of the Splunk index where logs are written. :type index: str, optional - :param indexed_fields: List of log field names to send as indexed fields to Splunk HEC. Available only when ``encoding`` is ``json``. - :type indexed_fields: [str], optional - :param inputs: A list of component IDs whose output is used as the ``input`` for this component. :type inputs: [str] @@ -141,8 +135,6 @@ def __init__( kwargs["endpoint_url_key"] = endpoint_url_key if index is not unset: kwargs["index"] = index - if indexed_fields is not unset: - kwargs["indexed_fields"] = indexed_fields if sourcetype is not unset: kwargs["sourcetype"] = sourcetype if token_key is not unset: diff --git a/src/datadog_api_client/v2/models/__init__.py b/src/datadog_api_client/v2/models/__init__.py index fe68361648..ee3b7d0034 100644 --- a/src/datadog_api_client/v2/models/__init__.py +++ b/src/datadog_api_client/v2/models/__init__.py @@ -3496,6 +3496,51 @@ from datadog_api_client.v2.model.observability_pipeline_amazon_s3_destination_type import ( ObservabilityPipelineAmazonS3DestinationType, ) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_batch_settings import ( + ObservabilityPipelineAmazonS3GenericBatchSettings, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression import ( + ObservabilityPipelineAmazonS3GenericCompression, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip import ( + ObservabilityPipelineAmazonS3GenericCompressionGzip, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_gzip_type import ( + ObservabilityPipelineAmazonS3GenericCompressionGzipType, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappy, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_snappy_type import ( + ObservabilityPipelineAmazonS3GenericCompressionSnappyType, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd import ( + ObservabilityPipelineAmazonS3GenericCompressionZstd, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_compression_zstd_type import ( + ObservabilityPipelineAmazonS3GenericCompressionZstdType, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination import ( + ObservabilityPipelineAmazonS3GenericDestination, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_destination_type import ( + ObservabilityPipelineAmazonS3GenericDestinationType, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding import ( + ObservabilityPipelineAmazonS3GenericEncoding, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json import ( + ObservabilityPipelineAmazonS3GenericEncodingJson, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_json_type import ( + ObservabilityPipelineAmazonS3GenericEncodingJsonType, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet import ( + ObservabilityPipelineAmazonS3GenericEncodingParquet, +) +from datadog_api_client.v2.model.observability_pipeline_amazon_s3_generic_encoding_parquet_type import ( + ObservabilityPipelineAmazonS3GenericEncodingParquetType, +) from datadog_api_client.v2.model.observability_pipeline_amazon_s3_source import ObservabilityPipelineAmazonS3Source from datadog_api_client.v2.model.observability_pipeline_amazon_s3_source_type import ( ObservabilityPipelineAmazonS3SourceType, @@ -9076,6 +9121,21 @@ "ObservabilityPipelineAmazonS3Destination", "ObservabilityPipelineAmazonS3DestinationStorageClass", "ObservabilityPipelineAmazonS3DestinationType", + "ObservabilityPipelineAmazonS3GenericBatchSettings", + "ObservabilityPipelineAmazonS3GenericCompression", + "ObservabilityPipelineAmazonS3GenericCompressionGzip", + "ObservabilityPipelineAmazonS3GenericCompressionGzipType", + "ObservabilityPipelineAmazonS3GenericCompressionSnappy", + "ObservabilityPipelineAmazonS3GenericCompressionSnappyType", + "ObservabilityPipelineAmazonS3GenericCompressionZstd", + "ObservabilityPipelineAmazonS3GenericCompressionZstdType", + "ObservabilityPipelineAmazonS3GenericDestination", + "ObservabilityPipelineAmazonS3GenericDestinationType", + "ObservabilityPipelineAmazonS3GenericEncoding", + "ObservabilityPipelineAmazonS3GenericEncodingJson", + "ObservabilityPipelineAmazonS3GenericEncodingJsonType", + "ObservabilityPipelineAmazonS3GenericEncodingParquet", + "ObservabilityPipelineAmazonS3GenericEncodingParquetType", "ObservabilityPipelineAmazonS3Source", "ObservabilityPipelineAmazonS3SourceType", "ObservabilityPipelineAmazonSecurityLakeDestination",