mirror of
https://github.com/OMGeeky/google-apis-rs.git
synced 2026-02-23 15:49:49 +01:00
Update the JSON schemas
The command executed was as follows: rm -f .api.deps .cli.deps && FETCH_APIS=1 make update-json -j8
This commit is contained in:
@@ -31,6 +31,11 @@
|
||||
"description": "Regional Endpoint",
|
||||
"endpointUrl": "https://dataflow.europe-west3.rep.googleapis.com/",
|
||||
"location": "europe-west3"
|
||||
},
|
||||
{
|
||||
"description": "Regional Endpoint",
|
||||
"endpointUrl": "https://dataflow.europe-west9.rep.googleapis.com/",
|
||||
"location": "europe-west9"
|
||||
}
|
||||
],
|
||||
"fullyEncodeReservedExpansion": true,
|
||||
@@ -1918,7 +1923,7 @@
|
||||
],
|
||||
"parameters": {
|
||||
"dynamicTemplate.gcsPath": {
|
||||
"description": "Path to dynamic template spec file on Cloud Storage. The file must be a Json serialized DynamicTemplateFieSpec object.",
|
||||
"description": "Path to the dynamic template specification file on Cloud Storage. The file must be a JSON serialized `DynamicTemplateFileSpec` object.",
|
||||
"location": "query",
|
||||
"type": "string"
|
||||
},
|
||||
@@ -1928,7 +1933,7 @@
|
||||
"type": "string"
|
||||
},
|
||||
"gcsPath": {
|
||||
"description": "A Cloud Storage path to the template from which to create the job. Must be valid Cloud Storage URL, beginning with 'gs://'.",
|
||||
"description": "A Cloud Storage path to the template to use to create the job. Must be valid Cloud Storage URL, beginning with `gs://`.",
|
||||
"location": "query",
|
||||
"type": "string"
|
||||
},
|
||||
@@ -2133,7 +2138,7 @@
|
||||
],
|
||||
"parameters": {
|
||||
"dynamicTemplate.gcsPath": {
|
||||
"description": "Path to dynamic template spec file on Cloud Storage. The file must be a Json serialized DynamicTemplateFieSpec object.",
|
||||
"description": "Path to the dynamic template specification file on Cloud Storage. The file must be a JSON serialized `DynamicTemplateFileSpec` object.",
|
||||
"location": "query",
|
||||
"type": "string"
|
||||
},
|
||||
@@ -2143,7 +2148,7 @@
|
||||
"type": "string"
|
||||
},
|
||||
"gcsPath": {
|
||||
"description": "A Cloud Storage path to the template from which to create the job. Must be valid Cloud Storage URL, beginning with 'gs://'.",
|
||||
"description": "A Cloud Storage path to the template to use to create the job. Must be valid Cloud Storage URL, beginning with `gs://`.",
|
||||
"location": "query",
|
||||
"type": "string"
|
||||
},
|
||||
@@ -2182,7 +2187,7 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"revision": "20240225",
|
||||
"revision": "20240415",
|
||||
"rootUrl": "https://dataflow.googleapis.com/",
|
||||
"schemas": {
|
||||
"ApproximateProgress": {
|
||||
@@ -3201,7 +3206,7 @@
|
||||
"type": "string"
|
||||
},
|
||||
"streamingMode": {
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case.",
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case. For more information, see [Set the pipeline streaming mode](https://cloud.google.com/dataflow/docs/guides/streaming-modes).",
|
||||
"enum": [
|
||||
"STREAMING_MODE_UNSPECIFIED",
|
||||
"STREAMING_MODE_EXACTLY_ONCE",
|
||||
@@ -3539,7 +3544,7 @@
|
||||
"type": "string"
|
||||
},
|
||||
"streamingMode": {
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case.",
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case. For more information, see [Set the pipeline streaming mode](https://cloud.google.com/dataflow/docs/guides/streaming-modes).",
|
||||
"enum": [
|
||||
"STREAMING_MODE_UNSPECIFIED",
|
||||
"STREAMING_MODE_EXACTLY_ONCE",
|
||||
@@ -5413,7 +5418,7 @@
|
||||
"type": "object"
|
||||
},
|
||||
"RuntimeEnvironment": {
|
||||
"description": "The environment values to set at runtime.",
|
||||
"description": "The environment values to set at runtime. LINT.IfChange",
|
||||
"id": "RuntimeEnvironment",
|
||||
"properties": {
|
||||
"additionalExperiments": {
|
||||
@@ -5484,7 +5489,7 @@
|
||||
"type": "string"
|
||||
},
|
||||
"streamingMode": {
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case.",
|
||||
"description": "Optional. Specifies the Streaming Engine message processing guarantees. Reduces cost and latency but might result in duplicate messages committed to storage. Designed to run simple mapping streaming ETL jobs at the lowest cost. For example, Change Data Capture (CDC) to BigQuery is a canonical use case. For more information, see [Set the pipeline streaming mode](https://cloud.google.com/dataflow/docs/guides/streaming-modes).",
|
||||
"enum": [
|
||||
"STREAMING_MODE_UNSPECIFIED",
|
||||
"STREAMING_MODE_EXACTLY_ONCE",
|
||||
@@ -5553,7 +5558,7 @@
|
||||
"type": "integer"
|
||||
},
|
||||
"workerUtilizationHint": {
|
||||
"description": "Target worker utilization, compared against the aggregate utilization of the worker pool by autoscaler, to determine upscaling and downscaling when absent other constraints such as backlog.",
|
||||
"description": "Target worker utilization, compared against the aggregate utilization of the worker pool by autoscaler, to determine upscaling and downscaling when absent other constraints such as backlog. For more information, see [Update an existing pipeline](https://cloud.google.com/dataflow/docs/guides/updating-a-pipeline).",
|
||||
"format": "double",
|
||||
"type": "number"
|
||||
}
|
||||
@@ -6958,6 +6963,10 @@
|
||||
"description": "Metadata describing a template.",
|
||||
"id": "TemplateMetadata",
|
||||
"properties": {
|
||||
"defaultStreamingMode": {
|
||||
"description": "Optional. Indicates the default streaming mode for a streaming template. Only valid if both supports_at_least_once and supports_exactly_once are true. Possible values: UNSPECIFIED, EXACTLY_ONCE and AT_LEAST_ONCE",
|
||||
"type": "string"
|
||||
},
|
||||
"description": {
|
||||
"description": "Optional. A description of the template.",
|
||||
"type": "string"
|
||||
|
||||
Reference in New Issue
Block a user