From 65971c547587cb85e752b25a836e0a7ac41eee48 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Wed, 5 Jul 2023 15:16:34 -0400 Subject: [PATCH 1/2] feat: add `model` and `language_codes` fields in `RecognitionConfig` message + enable default `_` recognizer (#524) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * feat: add `model` and `language_codes` fields in `RecognitionConfig` message + enable default `_` recognizer Enables specifying `model` and `language_codes` in requests without having to specify them in the Recognizer (they can still be specified in the Recognizer in the `default_recognition_config` field). Also enables using the recognizer ID `_` to perform recognition without explicitly creating a Recognizer resource. The top-level `model` and `language_codes` fields are deprecated in favor of the new fields added in the `RecognitionConfig` message. The old fields continue to work. PiperOrigin-RevId: 545698919 Source-Link: https://github.com/googleapis/googleapis/commit/e73fc8ff3b5c3b025254219de1ca8d092dc17e48 Source-Link: https://github.com/googleapis/googleapis-gen/commit/b77dfdfe3927e409875b2d8913f2574fa00a05a5 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYjc3ZGZkZmUzOTI3ZTQwOTg3NWIyZDg5MTNmMjU3NGZhMDBhMDVhNSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md --------- Co-authored-by: Owl Bot --- .../speech_v2/services/speech/async_client.py | 42 +++++----- .../cloud/speech_v2/services/speech/client.py | 42 +++++----- google/cloud/speech_v2/types/cloud_speech.py | 80 ++++++++++++++----- ...ippet_metadata_google.cloud.speech.v1.json | 2 +- ...etadata_google.cloud.speech.v1p1beta1.json | 2 +- ...ippet_metadata_google.cloud.speech.v2.json | 58 +++++++------- ...enerated_speech_create_recognizer_async.py | 5 -- ...generated_speech_create_recognizer_sync.py | 5 -- ...enerated_speech_update_recognizer_async.py | 5 -- ...generated_speech_update_recognizer_sync.py | 5 -- tests/unit/gapic/speech_v2/test_speech.py | 8 ++ 11 files changed, 146 insertions(+), 108 deletions(-) diff --git a/google/cloud/speech_v2/services/speech/async_client.py b/google/cloud/speech_v2/services/speech/async_client.py index 65a9a464..1b60d93f 100644 --- a/google/cloud/speech_v2/services/speech/async_client.py +++ b/google/cloud/speech_v2/services/speech/async_client.py @@ -254,12 +254,7 @@ async def sample_create_recognizer(): client = speech_v2.SpeechAsyncClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.CreateRecognizerRequest( - recognizer=recognizer, parent="parent_value", ) @@ -624,12 +619,7 @@ async def sample_update_recognizer(): client = speech_v2.SpeechAsyncClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.UpdateRecognizerRequest( - recognizer=recognizer, ) # Make the request @@ -1030,6 +1020,8 @@ async def sample_recognize(): Required. The name of the Recognizer to use during recognition. The expected format is ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an + empty implicit Recognizer. This corresponds to the ``recognizer`` field on the ``request`` instance; if ``request`` is provided, this @@ -1213,17 +1205,26 @@ def request_generator(): [StreamingRecognize][google.cloud.speech.v2.Speech.StreamingRecognize] method. Multiple [StreamingRecognizeRequest][google.cloud.speech.v2.StreamingRecognizeRequest] - messages are sent. The first message must contain a + messages are sent in one call. + + If the [Recognizer][google.cloud.speech.v2.Recognizer] + referenced by + [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] + contains a fully specified request configuration then + the stream may only contain messages with only + [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] + set. + + Otherwise the first message must contain a [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] - and optionally a + and a [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message and must not contain + message that together fully specify the request + configuration and must not contain [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio]. - All subsequent messages must contain + All subsequent messages must only have [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] - and must not contain a - [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message. + set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. @@ -1367,8 +1368,11 @@ async def sample_batch_recognize(): [BatchRecognize][google.cloud.speech.v2.Speech.BatchRecognize] method. recognizer (:class:`str`): - Required. Resource name of the - recognizer to be used for ASR. + Required. The name of the Recognizer to use during + recognition. The expected format is + ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an + empty implicit Recognizer. This corresponds to the ``recognizer`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/speech_v2/services/speech/client.py b/google/cloud/speech_v2/services/speech/client.py index 28e9435f..c91d4f47 100644 --- a/google/cloud/speech_v2/services/speech/client.py +++ b/google/cloud/speech_v2/services/speech/client.py @@ -593,12 +593,7 @@ def sample_create_recognizer(): client = speech_v2.SpeechClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.CreateRecognizerRequest( - recognizer=recognizer, parent="parent_value", ) @@ -963,12 +958,7 @@ def sample_update_recognizer(): client = speech_v2.SpeechClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.UpdateRecognizerRequest( - recognizer=recognizer, ) # Make the request @@ -1369,6 +1359,8 @@ def sample_recognize(): Required. The name of the Recognizer to use during recognition. The expected format is ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an + empty implicit Recognizer. This corresponds to the ``recognizer`` field on the ``request`` instance; if ``request`` is provided, this @@ -1550,17 +1542,26 @@ def request_generator(): [StreamingRecognize][google.cloud.speech.v2.Speech.StreamingRecognize] method. Multiple [StreamingRecognizeRequest][google.cloud.speech.v2.StreamingRecognizeRequest] - messages are sent. The first message must contain a + messages are sent in one call. + + If the [Recognizer][google.cloud.speech.v2.Recognizer] + referenced by + [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] + contains a fully specified request configuration then + the stream may only contain messages with only + [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] + set. + + Otherwise the first message must contain a [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] - and optionally a + and a [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message and must not contain + message that together fully specify the request + configuration and must not contain [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio]. - All subsequent messages must contain + All subsequent messages must only have [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] - and must not contain a - [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message. + set. retry (google.api_core.retry.Retry): Designation of what errors, if any, should be retried. timeout (float): The timeout for this request. @@ -1700,8 +1701,11 @@ def sample_batch_recognize(): [BatchRecognize][google.cloud.speech.v2.Speech.BatchRecognize] method. recognizer (str): - Required. Resource name of the - recognizer to be used for ASR. + Required. The name of the Recognizer to use during + recognition. The expected format is + ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an + empty implicit Recognizer. This corresponds to the ``recognizer`` field on the ``request`` instance; if ``request`` is provided, this diff --git a/google/cloud/speech_v2/types/cloud_speech.py b/google/cloud/speech_v2/types/cloud_speech.py index 007f4bd7..a659dfc0 100644 --- a/google/cloud/speech_v2/types/cloud_speech.py +++ b/google/cloud/speech_v2/types/cloud_speech.py @@ -589,7 +589,7 @@ class Recognizer(proto.Message): User-settable, human-readable name for the Recognizer. Must be 63 characters or less. model (str): - Required. Which model to use for recognition requests. + Optional. Which model to use for recognition requests. Select the model best suited to your domain to get best results. @@ -600,7 +600,7 @@ class Recognizer(proto.Message): `Table Of Supported Models `__. language_codes (MutableSequence[str]): - Required. The language of the supplied audio as a + Optional. The language of the supplied audio as a `BCP-47 `__ language tag. @@ -1080,6 +1080,31 @@ class RecognitionConfig(proto.Message): (linear16, mulaw, alaw). This field is a member of `oneof`_ ``decoding_config``. + model (str): + Optional. Which model to use for recognition requests. + Select the model best suited to your domain to get best + results. + + Guidance for choosing which model to use can be found in the + `Transcription Models + Documentation `__ + and the models supported in each region can be found in the + `Table Of Supported + Models `__. + language_codes (MutableSequence[str]): + Optional. The language of the supplied audio as a + `BCP-47 `__ + language tag. Language tags are normalized to BCP-47 before + they are used eg "en-us" becomes "en-US". + + Supported languages for each model are listed in the `Table + of Supported + Models `__. + + If additional languages are provided, recognition result + will contain recognition in the most likely language + detected. The recognition result will include the language + tag of the language detected in the audio. features (google.cloud.speech_v2.types.RecognitionFeatures): Speech recognition features to enable. adaptation (google.cloud.speech_v2.types.SpeechAdaptation): @@ -1100,6 +1125,14 @@ class RecognitionConfig(proto.Message): oneof="decoding_config", message="ExplicitDecodingConfig", ) + model: str = proto.Field( + proto.STRING, + number=9, + ) + language_codes: MutableSequence[str] = proto.RepeatedField( + proto.STRING, + number=10, + ) features: "RecognitionFeatures" = proto.Field( proto.MESSAGE, number=2, @@ -1132,6 +1165,8 @@ class RecognizeRequest(proto.Message): Required. The name of the Recognizer to use during recognition. The expected format is ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an empty + implicit Recognizer. config (google.cloud.speech_v2.types.RecognitionConfig): Features and audio metadata to use for the Automatic Speech Recognition. This field in combination with the @@ -1515,17 +1550,23 @@ class StreamingRecognizeRequest(proto.Message): [StreamingRecognize][google.cloud.speech.v2.Speech.StreamingRecognize] method. Multiple [StreamingRecognizeRequest][google.cloud.speech.v2.StreamingRecognizeRequest] - messages are sent. The first message must contain a + messages are sent in one call. + + If the [Recognizer][google.cloud.speech.v2.Recognizer] referenced by [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] - and optionally a + contains a fully specified request configuration then the stream may + only contain messages with only + [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] set. + + Otherwise the first message must contain a + [recognizer][google.cloud.speech.v2.StreamingRecognizeRequest.recognizer] + and a [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message and must not contain + message that together fully specify the request configuration and + must not contain [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio]. All - subsequent messages must contain - [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] and - must not contain a - [streaming_config][google.cloud.speech.v2.StreamingRecognizeRequest.streaming_config] - message. + subsequent messages must only have + [audio][google.cloud.speech.v2.StreamingRecognizeRequest.audio] set. This message has `oneof`_ fields (mutually exclusive fields). For each oneof, at most one member field can be set at the same time. @@ -1536,13 +1577,11 @@ class StreamingRecognizeRequest(proto.Message): Attributes: recognizer (str): - Required. Streaming recognition should start with an initial - request having a ``recognizer``. Subsequent requests carry - the audio data to be recognized. - - The initial request with configuration can be omitted if the - Recognizer being used has a - [default_recognition_config][google.cloud.speech.v2.Recognizer.default_recognition_config]. + Required. The name of the Recognizer to use during + recognition. The expected format is + ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an empty + implicit Recognizer. streaming_config (google.cloud.speech_v2.types.StreamingRecognitionConfig): StreamingRecognitionConfig to be used in this recognition attempt. If provided, it will @@ -1582,8 +1621,11 @@ class BatchRecognizeRequest(proto.Message): Attributes: recognizer (str): - Required. Resource name of the recognizer to - be used for ASR. + Required. The name of the Recognizer to use during + recognition. The expected format is + ``projects/{project}/locations/{location}/recognizers/{recognizer}``. + The {recognizer} segment may be set to ``_`` to use an empty + implicit Recognizer. config (google.cloud.speech_v2.types.RecognitionConfig): Features and audio metadata to use for the Automatic Speech Recognition. This field in combination with the diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json index 1b92cfc5..c793efe0 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "2.20.1" + "version": "0.1.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json index f37c1def..b51d437c 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "2.20.1" + "version": "0.1.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json index 054a164e..73ac2fc1 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "2.20.1" + "version": "0.1.0" }, "snippets": [ { @@ -607,12 +607,12 @@ "regionTag": "speech_v2_generated_Speech_CreateRecognizer_async", "segments": [ { - "end": 60, + "end": 55, "start": 27, "type": "FULL" }, { - "end": 60, + "end": 55, "start": 27, "type": "SHORT" }, @@ -622,18 +622,18 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 50, + "end": 45, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 57, - "start": 51, + "end": 52, + "start": 46, "type": "REQUEST_EXECUTION" }, { - "end": 61, - "start": 58, + "end": 56, + "start": 53, "type": "RESPONSE_HANDLING" } ], @@ -695,12 +695,12 @@ "regionTag": "speech_v2_generated_Speech_CreateRecognizer_sync", "segments": [ { - "end": 60, + "end": 55, "start": 27, "type": "FULL" }, { - "end": 60, + "end": 55, "start": 27, "type": "SHORT" }, @@ -710,18 +710,18 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 50, + "end": 45, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 57, - "start": 51, + "end": 52, + "start": 46, "type": "REQUEST_EXECUTION" }, { - "end": 61, - "start": 58, + "end": 56, + "start": 53, "type": "RESPONSE_HANDLING" } ], @@ -3726,12 +3726,12 @@ "regionTag": "speech_v2_generated_Speech_UpdateRecognizer_async", "segments": [ { - "end": 59, + "end": 54, "start": 27, "type": "FULL" }, { - "end": 59, + "end": 54, "start": 27, "type": "SHORT" }, @@ -3741,18 +3741,18 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 49, + "end": 44, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 56, - "start": 50, + "end": 51, + "start": 45, "type": "REQUEST_EXECUTION" }, { - "end": 60, - "start": 57, + "end": 55, + "start": 52, "type": "RESPONSE_HANDLING" } ], @@ -3810,12 +3810,12 @@ "regionTag": "speech_v2_generated_Speech_UpdateRecognizer_sync", "segments": [ { - "end": 59, + "end": 54, "start": 27, "type": "FULL" }, { - "end": 59, + "end": 54, "start": 27, "type": "SHORT" }, @@ -3825,18 +3825,18 @@ "type": "CLIENT_INITIALIZATION" }, { - "end": 49, + "end": 44, "start": 41, "type": "REQUEST_INITIALIZATION" }, { - "end": 56, - "start": 50, + "end": 51, + "start": 45, "type": "REQUEST_EXECUTION" }, { - "end": 60, - "start": 57, + "end": 55, + "start": 52, "type": "RESPONSE_HANDLING" } ], diff --git a/samples/generated_samples/speech_v2_generated_speech_create_recognizer_async.py b/samples/generated_samples/speech_v2_generated_speech_create_recognizer_async.py index 03dbbce7..fca73a0d 100644 --- a/samples/generated_samples/speech_v2_generated_speech_create_recognizer_async.py +++ b/samples/generated_samples/speech_v2_generated_speech_create_recognizer_async.py @@ -39,12 +39,7 @@ async def sample_create_recognizer(): client = speech_v2.SpeechAsyncClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.CreateRecognizerRequest( - recognizer=recognizer, parent="parent_value", ) diff --git a/samples/generated_samples/speech_v2_generated_speech_create_recognizer_sync.py b/samples/generated_samples/speech_v2_generated_speech_create_recognizer_sync.py index af0ffdc7..19374c37 100644 --- a/samples/generated_samples/speech_v2_generated_speech_create_recognizer_sync.py +++ b/samples/generated_samples/speech_v2_generated_speech_create_recognizer_sync.py @@ -39,12 +39,7 @@ def sample_create_recognizer(): client = speech_v2.SpeechClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.CreateRecognizerRequest( - recognizer=recognizer, parent="parent_value", ) diff --git a/samples/generated_samples/speech_v2_generated_speech_update_recognizer_async.py b/samples/generated_samples/speech_v2_generated_speech_update_recognizer_async.py index cc2f894c..8e044e81 100644 --- a/samples/generated_samples/speech_v2_generated_speech_update_recognizer_async.py +++ b/samples/generated_samples/speech_v2_generated_speech_update_recognizer_async.py @@ -39,12 +39,7 @@ async def sample_update_recognizer(): client = speech_v2.SpeechAsyncClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.UpdateRecognizerRequest( - recognizer=recognizer, ) # Make the request diff --git a/samples/generated_samples/speech_v2_generated_speech_update_recognizer_sync.py b/samples/generated_samples/speech_v2_generated_speech_update_recognizer_sync.py index 2c38e4fc..b7aa7615 100644 --- a/samples/generated_samples/speech_v2_generated_speech_update_recognizer_sync.py +++ b/samples/generated_samples/speech_v2_generated_speech_update_recognizer_sync.py @@ -39,12 +39,7 @@ def sample_update_recognizer(): client = speech_v2.SpeechClient() # Initialize request argument(s) - recognizer = speech_v2.Recognizer() - recognizer.model = "model_value" - recognizer.language_codes = ['language_codes_value1', 'language_codes_value2'] - request = speech_v2.UpdateRecognizerRequest( - recognizer=recognizer, ) # Make the request diff --git a/tests/unit/gapic/speech_v2/test_speech.py b/tests/unit/gapic/speech_v2/test_speech.py index a8a29bc7..432f4a52 100644 --- a/tests/unit/gapic/speech_v2/test_speech.py +++ b/tests/unit/gapic/speech_v2/test_speech.py @@ -6810,6 +6810,8 @@ def test_create_recognizer_rest(request_type): "sample_rate_hertz": 1817, "audio_channel_count": 2002, }, + "model": "model_value", + "language_codes": ["language_codes_value1", "language_codes_value2"], "features": { "profanity_filter": True, "enable_word_time_offsets": True, @@ -7083,6 +7085,8 @@ def test_create_recognizer_rest_bad_request( "sample_rate_hertz": 1817, "audio_channel_count": 2002, }, + "model": "model_value", + "language_codes": ["language_codes_value1", "language_codes_value2"], "features": { "profanity_filter": True, "enable_word_time_offsets": True, @@ -7886,6 +7890,8 @@ def test_update_recognizer_rest(request_type): "sample_rate_hertz": 1817, "audio_channel_count": 2002, }, + "model": "model_value", + "language_codes": ["language_codes_value1", "language_codes_value2"], "features": { "profanity_filter": True, "enable_word_time_offsets": True, @@ -8151,6 +8157,8 @@ def test_update_recognizer_rest_bad_request( "sample_rate_hertz": 1817, "audio_channel_count": 2002, }, + "model": "model_value", + "language_codes": ["language_codes_value1", "language_codes_value2"], "features": { "profanity_filter": True, "enable_word_time_offsets": True, From 6ca5097b8cecac11c341ad3601de07f4dff40ae0 Mon Sep 17 00:00:00 2001 From: "release-please[bot]" <55107282+release-please[bot]@users.noreply.github.com> Date: Wed, 5 Jul 2023 15:25:58 -0400 Subject: [PATCH 2/2] chore(main): release 2.21.0 (#525) Co-authored-by: release-please[bot] <55107282+release-please[bot]@users.noreply.github.com> --- .release-please-manifest.json | 2 +- CHANGELOG.md | 7 +++++++ google/cloud/speech/gapic_version.py | 2 +- google/cloud/speech_v1/gapic_version.py | 2 +- google/cloud/speech_v1p1beta1/gapic_version.py | 2 +- google/cloud/speech_v2/gapic_version.py | 2 +- .../snippet_metadata_google.cloud.speech.v1.json | 2 +- .../snippet_metadata_google.cloud.speech.v1p1beta1.json | 2 +- .../snippet_metadata_google.cloud.speech.v2.json | 2 +- 9 files changed, 15 insertions(+), 8 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index f2db52b4..1ff4c05f 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,4 +1,4 @@ { - ".": "2.20.1" + ".": "2.21.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index 420b9da6..b2b6e448 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -4,6 +4,13 @@ [1]: https://pypi.org/project/google-cloud-speech/#history +## [2.21.0](https://github.com/googleapis/python-speech/compare/v2.20.1...v2.21.0) (2023-07-05) + + +### Features + +* Add `model` and `language_codes` fields in `RecognitionConfig` message + enable default `_` recognizer ([#524](https://github.com/googleapis/python-speech/issues/524)) ([65971c5](https://github.com/googleapis/python-speech/commit/65971c547587cb85e752b25a836e0a7ac41eee48)) + ## [2.20.1](https://github.com/googleapis/python-speech/compare/v2.20.0...v2.20.1) (2023-07-04) diff --git a/google/cloud/speech/gapic_version.py b/google/cloud/speech/gapic_version.py index 5585b0b1..e546bae0 100644 --- a/google/cloud/speech/gapic_version.py +++ b/google/cloud/speech/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.20.1" # {x-release-please-version} +__version__ = "2.21.0" # {x-release-please-version} diff --git a/google/cloud/speech_v1/gapic_version.py b/google/cloud/speech_v1/gapic_version.py index 5585b0b1..e546bae0 100644 --- a/google/cloud/speech_v1/gapic_version.py +++ b/google/cloud/speech_v1/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.20.1" # {x-release-please-version} +__version__ = "2.21.0" # {x-release-please-version} diff --git a/google/cloud/speech_v1p1beta1/gapic_version.py b/google/cloud/speech_v1p1beta1/gapic_version.py index 5585b0b1..e546bae0 100644 --- a/google/cloud/speech_v1p1beta1/gapic_version.py +++ b/google/cloud/speech_v1p1beta1/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.20.1" # {x-release-please-version} +__version__ = "2.21.0" # {x-release-please-version} diff --git a/google/cloud/speech_v2/gapic_version.py b/google/cloud/speech_v2/gapic_version.py index 5585b0b1..e546bae0 100644 --- a/google/cloud/speech_v2/gapic_version.py +++ b/google/cloud/speech_v2/gapic_version.py @@ -13,4 +13,4 @@ # See the License for the specific language governing permissions and # limitations under the License. # -__version__ = "2.20.1" # {x-release-please-version} +__version__ = "2.21.0" # {x-release-please-version} diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json index c793efe0..58dd1792 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "0.1.0" + "version": "2.21.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json index b51d437c..25efd8af 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v1p1beta1.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "0.1.0" + "version": "2.21.0" }, "snippets": [ { diff --git a/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json b/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json index 73ac2fc1..bbf9aaa7 100644 --- a/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json +++ b/samples/generated_samples/snippet_metadata_google.cloud.speech.v2.json @@ -8,7 +8,7 @@ ], "language": "PYTHON", "name": "google-cloud-speech", - "version": "0.1.0" + "version": "2.21.0" }, "snippets": [ {