mirror of
https://github.com/OMGeeky/google-apis-rs.git
synced 2026-01-06 19:45:20 +01:00
update all APIs
Like documented in the README ``` rm -f .api.deps .cli.deps && FETCH_APIS=1 make update-json -j8 ```
This commit is contained in:
@@ -133,7 +133,7 @@
|
||||
]
|
||||
},
|
||||
"list": {
|
||||
"description": "Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `\"/v1/{name=users/*}/operations\"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.",
|
||||
"description": "Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`.",
|
||||
"flatPath": "v1/operations",
|
||||
"httpMethod": "GET",
|
||||
"id": "speech.operations.list",
|
||||
@@ -339,7 +339,7 @@
|
||||
],
|
||||
"parameters": {
|
||||
"parent": {
|
||||
"description": "Required. The parent resource where this phrase set will be created. Format: `projects/{project}/locations/{location}/phraseSets` Speech-to-Text supports three locations: `global`, `us` (US North America), and `eu` (Europe). If you are calling the `speech.googleapis.com` endpoint, use the `global` location. To specify a region, use a [regional endpoint](https://cloud.google.com/speech-to-text/docs/endpoints) with matching `us` or `eu` location value.",
|
||||
"description": "Required. The parent resource where this phrase set will be created. Format: `projects/{project}/locations/{location}` Speech-to-Text supports three locations: `global`, `us` (US North America), and `eu` (Europe). If you are calling the `speech.googleapis.com` endpoint, use the `global` location. To specify a region, use a [regional endpoint](https://cloud.google.com/speech-to-text/docs/endpoints) with matching `us` or `eu` location value.",
|
||||
"location": "path",
|
||||
"pattern": "^projects/[^/]+/locations/[^/]+$",
|
||||
"required": true,
|
||||
@@ -524,7 +524,7 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"revision": "20230119",
|
||||
"revision": "20240222",
|
||||
"rootUrl": "https://speech.googleapis.com/",
|
||||
"schemas": {
|
||||
"ABNFGrammar": {
|
||||
@@ -585,10 +585,40 @@
|
||||
"description": "A set of words or phrases that represents a common concept likely to appear in your audio, for example a list of passenger ship names. CustomClass items can be substituted into placeholders that you set in PhraseSet phrases.",
|
||||
"id": "CustomClass",
|
||||
"properties": {
|
||||
"annotations": {
|
||||
"additionalProperties": {
|
||||
"type": "string"
|
||||
},
|
||||
"description": "Output only. Allows users to store small amounts of arbitrary data. Both the key and the value must be 63 characters or less each. At most 100 annotations. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "object"
|
||||
},
|
||||
"customClassId": {
|
||||
"description": "If this custom class is a resource, the custom_class_id is the resource id of the CustomClass. Case sensitive.",
|
||||
"type": "string"
|
||||
},
|
||||
"deleteTime": {
|
||||
"description": "Output only. The time at which this resource was requested for deletion. This field is not used.",
|
||||
"format": "google-datetime",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"displayName": {
|
||||
"description": "Output only. User-settable, human-readable name for the CustomClass. Must be 63 characters or less. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"etag": {
|
||||
"description": "Output only. This checksum is computed by the server based on the value of other fields. This may be sent on update, undelete, and delete requests to ensure the client has an up-to-date value before proceeding. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"expireTime": {
|
||||
"description": "Output only. The time at which this resource will be purged. This field is not used.",
|
||||
"format": "google-datetime",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"items": {
|
||||
"description": "A collection of class items.",
|
||||
"items": {
|
||||
@@ -596,9 +626,44 @@
|
||||
},
|
||||
"type": "array"
|
||||
},
|
||||
"kmsKeyName": {
|
||||
"description": "Output only. The [KMS key name](https://cloud.google.com/kms/docs/resource-hierarchy#keys) with which the content of the ClassItem is encrypted. The expected format is `projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}`.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"kmsKeyVersionName": {
|
||||
"description": "Output only. The [KMS key version name](https://cloud.google.com/kms/docs/resource-hierarchy#key_versions) with which content of the ClassItem is encrypted. The expected format is `projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}/cryptoKeyVersions/{crypto_key_version}`.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"name": {
|
||||
"description": "The resource name of the custom class.",
|
||||
"type": "string"
|
||||
},
|
||||
"reconciling": {
|
||||
"description": "Output only. Whether or not this CustomClass is in the process of being updated. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "boolean"
|
||||
},
|
||||
"state": {
|
||||
"description": "Output only. The CustomClass lifecycle state. This field is not used.",
|
||||
"enum": [
|
||||
"STATE_UNSPECIFIED",
|
||||
"ACTIVE",
|
||||
"DELETED"
|
||||
],
|
||||
"enumDescriptions": [
|
||||
"Unspecified state. This is only used/useful for distinguishing unset values.",
|
||||
"The normal and active state.",
|
||||
"This CustomClass has been deleted."
|
||||
],
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"uid": {
|
||||
"description": "Output only. System-assigned unique identifier for the CustomClass. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
}
|
||||
},
|
||||
"type": "object"
|
||||
@@ -609,6 +674,25 @@
|
||||
"properties": {},
|
||||
"type": "object"
|
||||
},
|
||||
"Entry": {
|
||||
"description": "A single replacement configuration.",
|
||||
"id": "Entry",
|
||||
"properties": {
|
||||
"caseSensitive": {
|
||||
"description": "Whether the search is case sensitive.",
|
||||
"type": "boolean"
|
||||
},
|
||||
"replace": {
|
||||
"description": "What to replace with. Max length is 100 characters.",
|
||||
"type": "string"
|
||||
},
|
||||
"search": {
|
||||
"description": "What to replace. Max length is 100 characters.",
|
||||
"type": "string"
|
||||
}
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"ListCustomClassesResponse": {
|
||||
"description": "Message returned to the client by the `ListCustomClasses` method.",
|
||||
"id": "ListCustomClassesResponse",
|
||||
@@ -774,7 +858,7 @@
|
||||
"description": "Properties of the object. Contains field @type with type URL.",
|
||||
"type": "any"
|
||||
},
|
||||
"description": "The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.",
|
||||
"description": "The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.",
|
||||
"type": "object"
|
||||
}
|
||||
},
|
||||
@@ -800,11 +884,51 @@
|
||||
"description": "Provides \"hints\" to the speech recognizer to favor specific words and phrases in the results.",
|
||||
"id": "PhraseSet",
|
||||
"properties": {
|
||||
"annotations": {
|
||||
"additionalProperties": {
|
||||
"type": "string"
|
||||
},
|
||||
"description": "Output only. Allows users to store small amounts of arbitrary data. Both the key and the value must be 63 characters or less each. At most 100 annotations. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "object"
|
||||
},
|
||||
"boost": {
|
||||
"description": "Hint Boost. Positive value will increase the probability that a specific phrase will be recognized over other similar sounding phrases. The higher the boost, the higher the chance of false positive recognition as well. Negative boost values would correspond to anti-biasing. Anti-biasing is not enabled, so negative boost will simply be ignored. Though `boost` can accept a wide range of positive values, most use cases are best served with values between 0 (exclusive) and 20. We recommend using a binary search approach to finding the optimal value for your use case as well as adding phrases both with and without boost to your requests.",
|
||||
"format": "float",
|
||||
"type": "number"
|
||||
},
|
||||
"deleteTime": {
|
||||
"description": "Output only. The time at which this resource was requested for deletion. This field is not used.",
|
||||
"format": "google-datetime",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"displayName": {
|
||||
"description": "Output only. User-settable, human-readable name for the PhraseSet. Must be 63 characters or less. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"etag": {
|
||||
"description": "Output only. This checksum is computed by the server based on the value of other fields. This may be sent on update, undelete, and delete requests to ensure the client has an up-to-date value before proceeding. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"expireTime": {
|
||||
"description": "Output only. The time at which this resource will be purged. This field is not used.",
|
||||
"format": "google-datetime",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"kmsKeyName": {
|
||||
"description": "Output only. The [KMS key name](https://cloud.google.com/kms/docs/resource-hierarchy#keys) with which the content of the PhraseSet is encrypted. The expected format is `projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}`.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"kmsKeyVersionName": {
|
||||
"description": "Output only. The [KMS key version name](https://cloud.google.com/kms/docs/resource-hierarchy#key_versions) with which content of the PhraseSet is encrypted. The expected format is `projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}/cryptoKeyVersions/{crypto_key_version}`.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"name": {
|
||||
"description": "The resource name of the phrase set.",
|
||||
"type": "string"
|
||||
@@ -815,6 +939,31 @@
|
||||
"$ref": "Phrase"
|
||||
},
|
||||
"type": "array"
|
||||
},
|
||||
"reconciling": {
|
||||
"description": "Output only. Whether or not this PhraseSet is in the process of being updated. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "boolean"
|
||||
},
|
||||
"state": {
|
||||
"description": "Output only. The CustomClass lifecycle state. This field is not used.",
|
||||
"enum": [
|
||||
"STATE_UNSPECIFIED",
|
||||
"ACTIVE",
|
||||
"DELETED"
|
||||
],
|
||||
"enumDescriptions": [
|
||||
"Unspecified state. This is only used/useful for distinguishing unset values.",
|
||||
"The normal and active state.",
|
||||
"This CustomClass has been deleted."
|
||||
],
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"uid": {
|
||||
"description": "Output only. System-assigned unique identifier for the PhraseSet. This field is not used.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
}
|
||||
},
|
||||
"type": "object"
|
||||
@@ -894,6 +1043,7 @@
|
||||
"AMR_WB",
|
||||
"OGG_OPUS",
|
||||
"SPEEX_WITH_HEADER_BYTE",
|
||||
"MP3",
|
||||
"WEBM_OPUS"
|
||||
],
|
||||
"enumDescriptions": [
|
||||
@@ -905,7 +1055,8 @@
|
||||
"Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000.",
|
||||
"Opus encoded audio frames in Ogg container ([OggOpus](https://wiki.xiph.org/OggOpus)). `sample_rate_hertz` must be one of 8000, 12000, 16000, 24000, or 48000.",
|
||||
"Although the use of lossy encodings is not recommended, if a very low bitrate encoding is required, `OGG_OPUS` is highly preferred over Speex encoding. The [Speex](https://speex.org/) encoding supported by Cloud Speech API has a header byte in each block, as in MIME type `audio/x-speex-with-header-byte`. It is a variant of the RTP Speex encoding defined in [RFC 5574](https://tools.ietf.org/html/rfc5574). The stream is a sequence of blocks, one block per RTP packet. Each block starts with a byte containing the length of the block, in bytes, followed by one or more frames of Speex data, padded to an integral number of bytes (octets) as specified in RFC 5574. In other words, each RTP header is replaced with a single byte containing the block length. Only Speex wideband is supported. `sample_rate_hertz` must be 16000.",
|
||||
"Opus encoded audio frames in WebM container ([OggOpus](https://wiki.xiph.org/OggOpus)). `sample_rate_hertz` must be one of 8000, 12000, 16000, 24000, or 48000."
|
||||
"MP3 audio. MP3 encoding is a Beta feature and only available in v1p1beta1. Support all standard MP3 bitrates (which range from 32-320 kbps). When using this encoding, `sample_rate_hertz` has to match the sample rate of the file being used.",
|
||||
"Opus encoded audio frames in WebM container ([WebM](https://www.webmproject.org/docs/container/)). `sample_rate_hertz` must be one of 8000, 12000, 16000, 24000, or 48000."
|
||||
],
|
||||
"type": "string"
|
||||
},
|
||||
@@ -942,6 +1093,10 @@
|
||||
},
|
||||
"type": "array"
|
||||
},
|
||||
"transcriptNormalization": {
|
||||
"$ref": "TranscriptNormalization",
|
||||
"description": "Optional. Use transcription normalization to automatically replace parts of the transcript with phrases of your choosing. For StreamingRecognize, this normalization only applies to stable partial transcripts (stability > 0.8) and final transcripts."
|
||||
},
|
||||
"useEnhanced": {
|
||||
"description": "Set to true to use an enhanced model for speech recognition. If `use_enhanced` is set to true and the `model` field is not set, then an appropriate enhanced model is chosen if an enhanced model exists for the audio. If `use_enhanced` is true and an enhanced version of the specified model does not exist, then the speech is recognized using the standard version of the specified model.",
|
||||
"type": "boolean"
|
||||
@@ -950,6 +1105,7 @@
|
||||
"type": "object"
|
||||
},
|
||||
"RecognitionMetadata": {
|
||||
"deprecated": true,
|
||||
"description": "Description of audio data to be recognized.",
|
||||
"id": "RecognitionMetadata",
|
||||
"properties": {
|
||||
@@ -1090,6 +1246,10 @@
|
||||
"description": "When available, billed audio seconds for the corresponding request.",
|
||||
"format": "google-duration",
|
||||
"type": "string"
|
||||
},
|
||||
"usingLegacyModels": {
|
||||
"description": "Whether request used legacy asr models (was not automatically migrated to use conformer models).",
|
||||
"type": "boolean"
|
||||
}
|
||||
},
|
||||
"type": "object"
|
||||
@@ -1099,7 +1259,7 @@
|
||||
"id": "SpeakerDiarizationConfig",
|
||||
"properties": {
|
||||
"enableSpeakerDiarization": {
|
||||
"description": "If 'true', enables speaker detection for each recognized word in the top alternative of the recognition result using a speaker_tag provided in the WordInfo.",
|
||||
"description": "If 'true', enables speaker detection for each recognized word in the top alternative of the recognition result using a speaker_label provided in the WordInfo.",
|
||||
"type": "boolean"
|
||||
},
|
||||
"maxSpeakerCount": {
|
||||
@@ -1113,6 +1273,7 @@
|
||||
"type": "integer"
|
||||
},
|
||||
"speakerTag": {
|
||||
"deprecated": true,
|
||||
"description": "Output only. Unused.",
|
||||
"format": "int32",
|
||||
"readOnly": true,
|
||||
@@ -1266,6 +1427,20 @@
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"TranscriptNormalization": {
|
||||
"description": "Transcription normalization configuration. Use transcription normalization to automatically replace parts of the transcript with phrases of your choosing. For StreamingRecognize, this normalization only applies to stable partial transcripts (stability > 0.8) and final transcripts.",
|
||||
"id": "TranscriptNormalization",
|
||||
"properties": {
|
||||
"entries": {
|
||||
"description": "A list of replacement entries. We will perform replacement with one entry at a time. For example, the second entry in [\"cat\" => \"dog\", \"mountain cat\" => \"mountain dog\"] will never be applied because we will always process the first entry before it. At most 100 entries.",
|
||||
"items": {
|
||||
"$ref": "Entry"
|
||||
},
|
||||
"type": "array"
|
||||
}
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"TranscriptOutputConfig": {
|
||||
"description": "Specifies an optional destination for the recognition results.",
|
||||
"id": "TranscriptOutputConfig",
|
||||
@@ -1291,8 +1466,14 @@
|
||||
"format": "google-duration",
|
||||
"type": "string"
|
||||
},
|
||||
"speakerLabel": {
|
||||
"description": "Output only. A label value assigned for every unique speaker within the audio. This field specifies which speaker was detected to have spoken this word. For some models, like medical_conversation this can be actual speaker role, for example \"patient\" or \"provider\", but generally this would be a number identifying a speaker. This field is only set if enable_speaker_diarization = 'true' and only for the top alternative.",
|
||||
"readOnly": true,
|
||||
"type": "string"
|
||||
},
|
||||
"speakerTag": {
|
||||
"description": "Output only. A distinct integer value is assigned for every speaker within the audio. This field specifies which one of those speakers was detected to have spoken this word. Value ranges from '1' to diarization_speaker_count. speaker_tag is set if enable_speaker_diarization = 'true' and only in the top alternative.",
|
||||
"deprecated": true,
|
||||
"description": "Output only. A distinct integer value is assigned for every speaker within the audio. This field specifies which one of those speakers was detected to have spoken this word. Value ranges from '1' to diarization_speaker_count. speaker_tag is set if enable_speaker_diarization = 'true' and only for the top alternative. Note: Use speaker_label instead.",
|
||||
"format": "int32",
|
||||
"readOnly": true,
|
||||
"type": "integer"
|
||||
|
||||
Reference in New Issue
Block a user