From 5a09531b4c7d03c94426986d57378e50222a1b25 Mon Sep 17 00:00:00 2001 From: "ci.datadog-api-spec" Date: Wed, 11 Mar 2026 17:09:52 +0000 Subject: [PATCH] Regenerate client from commit 608090b of spec repo --- .generator/schemas/v2/openapi.yaml | 68 ++++++++++++++- .../ValidatePipeline_884022323.py | 84 ------------------- ...rvability_pipeline_opentelemetry_source.py | 2 +- ...elemetry_source_returns_ok_response.frozen | 1 - ...ntelemetry_source_returns_ok_response.yaml | 23 ----- .../features/observability_pipelines.feature | 8 -- 6 files changed, 67 insertions(+), 119 deletions(-) delete mode 100644 examples/v2/observability-pipelines/ValidatePipeline_884022323.py delete mode 100644 tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.frozen delete mode 100644 tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.yaml diff --git a/.generator/schemas/v2/openapi.yaml b/.generator/schemas/v2/openapi.yaml index 3bec741a12..9b0ed7bd48 100644 --- a/.generator/schemas/v2/openapi.yaml +++ b/.generator/schemas/v2/openapi.yaml @@ -7145,6 +7145,8 @@ components: example: - processor-id items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -39884,6 +39886,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array table: @@ -41857,6 +41861,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -41936,6 +41942,8 @@ components: example: - datadog-agent-source items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array key_prefix: @@ -42068,6 +42076,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array region: @@ -42167,6 +42177,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -42342,6 +42354,7 @@ components: example: - datadog-agent-source items: + description: The ID of a component whose output is used as input. type: string type: array processors: @@ -42442,6 +42455,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array tls: @@ -42670,6 +42685,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array routes: @@ -42741,6 +42758,8 @@ components: example: - metric-tags-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -42794,6 +42813,7 @@ components: - service - version items: + description: A Datadog tag key to include or exclude. type: string type: array mode: @@ -42872,6 +42892,7 @@ components: - log.message - log.error items: + description: A log field path to evaluate for duplicate values. type: string type: array id: @@ -42978,6 +42999,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -43244,6 +43267,7 @@ components: description: List of column names to include from the reference table. If not provided, all columns are included. items: + description: The name of a column to include from the reference table. type: string type: array key_field: @@ -43464,6 +43488,7 @@ components: - service - env items: + description: A log field name used to group the metric series. type: string type: array include: @@ -43567,6 +43592,8 @@ components: example: - parse-json-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array log_type: @@ -43631,6 +43658,8 @@ components: example: - datadog-agent-source items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array key_prefix: @@ -43724,6 +43753,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array project: @@ -43841,6 +43872,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array password_key: @@ -44102,6 +44135,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array key_field: @@ -44276,6 +44311,7 @@ components: - topic1 - topic2 items: + description: A Kafka topic name to subscribe to. type: string type: array type: @@ -44437,6 +44473,7 @@ components: - service - version items: + description: A metric tag key to include or exclude based on the action. type: string type: array mode: @@ -44502,6 +44539,8 @@ components: example: - parse-json-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array license_key_key: @@ -44710,6 +44749,7 @@ components: example: - container items: + description: The name of an OCSF profile to apply to the event. type: string type: array version: @@ -44779,6 +44819,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -44821,7 +44863,7 @@ components: Protocol (OTLP) over gRPC and HTTP. - **Supported pipeline types:** logs, metrics' + **Supported pipeline types:** logs' properties: grpc_address_key: description: Environment variable name containing the gRPC server address @@ -44851,7 +44893,6 @@ components: type: object x-pipeline-types: - logs - - metrics ObservabilityPipelineOpentelemetrySourceType: default: opentelemetry description: The source type. The value should always be `opentelemetry`. @@ -45161,6 +45202,7 @@ components: Quotas are tracked independently by unique combinations of these field values. items: + description: The name of a log field used to partition quota enforcement. type: string type: array too_many_buckets_action: @@ -45267,6 +45309,7 @@ components: - log.user.id - log.device.id items: + description: A log field path used to group events for aggregation. type: string type: array id: @@ -45366,6 +45409,7 @@ components: - field1 - field2 items: + description: The name of a field to remove from the log event. type: string type: array id: @@ -45494,6 +45538,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array keepalive: @@ -45582,6 +45628,7 @@ components: - service - host items: + description: A log field name used to group events for independent sampling. type: string minItems: 1 type: array @@ -45815,6 +45862,8 @@ components: - card - account items: + description: A keyword string that reinforces detection when found near + the sensitive pattern. type: string type: array proximity: @@ -45895,6 +45944,8 @@ components: - pii - ccn items: + description: A tag string used to classify and filter this sensitive data + rule. type: string type: array required: @@ -45974,6 +46025,8 @@ components: example: - '' items: + description: A log field path to include or exclude from sensitive data + scanning. type: string type: array required: @@ -46006,6 +46059,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array region: @@ -46076,6 +46131,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array mode: @@ -46468,6 +46525,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array sourcetype: @@ -46626,6 +46685,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array type: @@ -46731,6 +46792,8 @@ components: example: - filter-processor items: + description: The ID of a component whose output is used as input for this + destination. type: string type: array keepalive: @@ -46828,6 +46891,7 @@ components: example: - log.user.id items: + description: A log field name used to group events for independent throttling. type: string type: array id: diff --git a/examples/v2/observability-pipelines/ValidatePipeline_884022323.py b/examples/v2/observability-pipelines/ValidatePipeline_884022323.py deleted file mode 100644 index fda04b4c53..0000000000 --- a/examples/v2/observability-pipelines/ValidatePipeline_884022323.py +++ /dev/null @@ -1,84 +0,0 @@ -""" -Validate a metrics pipeline with opentelemetry source returns "OK" response -""" - -from datadog_api_client import ApiClient, Configuration -from datadog_api_client.v2.api.observability_pipelines_api import ObservabilityPipelinesApi -from datadog_api_client.v2.model.observability_pipeline_config import ObservabilityPipelineConfig -from datadog_api_client.v2.model.observability_pipeline_config_pipeline_type import ( - ObservabilityPipelineConfigPipelineType, -) -from datadog_api_client.v2.model.observability_pipeline_config_processor_group import ( - ObservabilityPipelineConfigProcessorGroup, -) -from datadog_api_client.v2.model.observability_pipeline_data_attributes import ObservabilityPipelineDataAttributes -from datadog_api_client.v2.model.observability_pipeline_datadog_metrics_destination import ( - ObservabilityPipelineDatadogMetricsDestination, -) -from datadog_api_client.v2.model.observability_pipeline_datadog_metrics_destination_type import ( - ObservabilityPipelineDatadogMetricsDestinationType, -) -from datadog_api_client.v2.model.observability_pipeline_filter_processor import ObservabilityPipelineFilterProcessor -from datadog_api_client.v2.model.observability_pipeline_filter_processor_type import ( - ObservabilityPipelineFilterProcessorType, -) -from datadog_api_client.v2.model.observability_pipeline_opentelemetry_source import ( - ObservabilityPipelineOpentelemetrySource, -) -from datadog_api_client.v2.model.observability_pipeline_opentelemetry_source_type import ( - ObservabilityPipelineOpentelemetrySourceType, -) -from datadog_api_client.v2.model.observability_pipeline_spec import ObservabilityPipelineSpec -from datadog_api_client.v2.model.observability_pipeline_spec_data import ObservabilityPipelineSpecData - -body = ObservabilityPipelineSpec( - data=ObservabilityPipelineSpecData( - attributes=ObservabilityPipelineDataAttributes( - config=ObservabilityPipelineConfig( - pipeline_type=ObservabilityPipelineConfigPipelineType.METRICS, - destinations=[ - ObservabilityPipelineDatadogMetricsDestination( - id="datadog-metrics-destination", - inputs=[ - "my-processor-group", - ], - type=ObservabilityPipelineDatadogMetricsDestinationType.DATADOG_METRICS, - ), - ], - processor_groups=[ - ObservabilityPipelineConfigProcessorGroup( - enabled=True, - id="my-processor-group", - include="*", - inputs=[ - "opentelemetry-source", - ], - processors=[ - ObservabilityPipelineFilterProcessor( - enabled=True, - id="filter-processor", - include="env:production", - type=ObservabilityPipelineFilterProcessorType.FILTER, - ), - ], - ), - ], - sources=[ - ObservabilityPipelineOpentelemetrySource( - id="opentelemetry-source", - type=ObservabilityPipelineOpentelemetrySourceType.OPENTELEMETRY, - ), - ], - ), - name="Metrics OTel Pipeline", - ), - type="pipelines", - ), -) - -configuration = Configuration() -with ApiClient(configuration) as api_client: - api_instance = ObservabilityPipelinesApi(api_client) - response = api_instance.validate_pipeline(body=body) - - print(response) diff --git a/src/datadog_api_client/v2/model/observability_pipeline_opentelemetry_source.py b/src/datadog_api_client/v2/model/observability_pipeline_opentelemetry_source.py index da09b968f1..ac6090dcd9 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_opentelemetry_source.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_opentelemetry_source.py @@ -56,7 +56,7 @@ def __init__( """ The ``opentelemetry`` source receives telemetry data using the OpenTelemetry Protocol (OTLP) over gRPC and HTTP. - **Supported pipeline types:** logs, metrics + **Supported pipeline types:** logs :param grpc_address_key: Environment variable name containing the gRPC server address for receiving OTLP data. Must be a valid environment variable name (alphanumeric characters and underscores only). :type grpc_address_key: str, optional diff --git a/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.frozen b/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.frozen deleted file mode 100644 index 9218d08a36..0000000000 --- a/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.frozen +++ /dev/null @@ -1 +0,0 @@ -2026-03-10T16:11:47.487Z \ No newline at end of file diff --git a/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.yaml b/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.yaml deleted file mode 100644 index b09e25bdb7..0000000000 --- a/tests/v2/cassettes/test_scenarios/test_validate_a_metrics_pipeline_with_opentelemetry_source_returns_ok_response.yaml +++ /dev/null @@ -1,23 +0,0 @@ -interactions: -- request: - body: '{"data":{"attributes":{"config":{"destinations":[{"id":"datadog-metrics-destination","inputs":["my-processor-group"],"type":"datadog_metrics"}],"pipeline_type":"metrics","processor_groups":[{"enabled":true,"id":"my-processor-group","include":"*","inputs":["opentelemetry-source"],"processors":[{"enabled":true,"id":"filter-processor","include":"env:production","type":"filter"}]}],"sources":[{"id":"opentelemetry-source","type":"opentelemetry"}]},"name":"Metrics - OTel Pipeline"},"type":"pipelines"}}' - headers: - accept: - - application/json - content-type: - - application/json - method: POST - uri: https://api.datadoghq.com/api/v2/obs-pipelines/pipelines/validate - response: - body: - string: '{"errors":[]} - - ' - headers: - content-type: - - application/vnd.api+json - status: - code: 200 - message: OK -version: 1 diff --git a/tests/v2/features/observability_pipelines.feature b/tests/v2/features/observability_pipelines.feature index e272303955..bf90c06c13 100644 --- a/tests/v2/features/observability_pipelines.feature +++ b/tests/v2/features/observability_pipelines.feature @@ -149,14 +149,6 @@ Feature: Observability Pipelines And the response "data.attributes.config.destinations" has length 1 And the response "data.attributes.config.destinations[0].id" is equal to "updated-datadog-logs-destination-id" - @team:DataDog/observability-pipelines - Scenario: Validate a metrics pipeline with opentelemetry source returns "OK" response - Given new "ValidatePipeline" request - And body with value {"data": {"attributes": {"config": {"pipeline_type": "metrics", "destinations": [{"id": "datadog-metrics-destination", "inputs": ["my-processor-group"], "type": "datadog_metrics"}], "processor_groups": [{"enabled": true, "id": "my-processor-group", "include": "*", "inputs": ["opentelemetry-source"], "processors": [{"enabled": true, "id": "filter-processor", "include": "env:production", "type": "filter"}]}], "sources": [{"id": "opentelemetry-source", "type": "opentelemetry"}]}, "name": "Metrics OTel Pipeline"}, "type": "pipelines"}} - When the request is sent - Then the response status is 200 OK - And the response "errors" has length 0 - @team:DataDog/observability-pipelines Scenario: Validate an observability pipeline returns "Bad Request" response Given new "ValidatePipeline" request