You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<pclass="firstline">Initializes a location-level encryption key specification. An error will be thrown if the location has resources already created before the initialization. Once the encryption specification is initialized at a location, it is immutable and all newly created resources under the location will be encrypted with the existing specification.</p>
<pre>Initializes a location-level encryption key specification. An error will be thrown if the location has resources already created before the initialization. Once the encryption specification is initialized at a location, it is immutable and all newly created resources under the location will be encrypted with the existing specification.
92
+
93
+
Args:
94
+
name: string, Immutable. The resource name of the encryption key specification resource. Format: projects/{project}/locations/{location}/encryptionSpec (required)
95
+
body: object, The request body.
96
+
The object takes the form of:
97
+
98
+
{ # The request to initialize a location-level encryption specification.
99
+
"encryptionSpec": { # A customer-managed encryption key specification that can be applied to all created resources (e.g. Conversation). # Required. The encryption spec used for CMEK encryption. It is required that the kms key is in the same region as the endpoint. The same key will be used for all provisioned resources, if encryption is available. If the kms_key_name is left empty, no encryption will be enforced.
100
+
"kmsKey": "A String", # Required. The name of customer-managed encryption key that is used to secure a resource and its sub-resources. If empty, the resource is secured by the default Google encryption key. Only the key in the same location as this resource is allowed to be used for encryption. Format: `projects/{project}/locations/{location}/keyRings/{keyRing}/cryptoKeys/{key}`
101
+
"name": "A String", # Immutable. The resource name of the encryption key specification resource. Format: projects/{project}/locations/{location}/encryptionSpec
102
+
},
103
+
}
104
+
105
+
x__xgafv: string, V1 error format.
106
+
Allowed values
107
+
1 - v1 error format
108
+
2 - v2 error format
109
+
110
+
Returns:
111
+
An object of the form:
112
+
113
+
{ # This resource represents a long-running operation that is the result of a network API call.
114
+
"done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
115
+
"error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
116
+
"code": 42, # The status code, which should be an enum value of google.rpc.Code.
117
+
"details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
118
+
{
119
+
"a_key": "", # Properties of the object. Contains field @type with type URL.
120
+
},
121
+
],
122
+
"message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
123
+
},
124
+
"metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
125
+
"a_key": "", # Properties of the object. Contains field @type with type URL.
126
+
},
127
+
"name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
128
+
"response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
129
+
"a_key": "", # Properties of the object. Contains field @type with type URL.
name: string, Required. The name of the encryption spec resource to get. (required)
250
+
x__xgafv: string, V1 error format.
251
+
Allowed values
252
+
1 - v1 error format
253
+
2 - v2 error format
254
+
255
+
Returns:
256
+
An object of the form:
257
+
258
+
{ # A customer-managed encryption key specification that can be applied to all created resources (e.g. Conversation).
259
+
"kmsKey": "A String", # Required. The name of customer-managed encryption key that is used to secure a resource and its sub-resources. If empty, the resource is secured by the default Google encryption key. Only the key in the same location as this resource is allowed to be used for encryption. Format: `projects/{project}/locations/{location}/keyRings/{keyRing}/cryptoKeys/{key}`
260
+
"name": "A String", # Immutable. The resource name of the encryption key specification resource. Format: projects/{project}/locations/{location}/encryptionSpec
Copy file name to clipboardExpand all lines: docs/dyn/dialogflow_v2.projects.locations.suggestions.html
+7Lines changed: 7 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -262,7 +262,14 @@ <h3>Method Details</h3>
262
262
},
263
263
"securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
264
264
"sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
265
+
"alternativeLanguageCodes": [ # Defines the list of other language codes in addition to the one provided by the conversation profile that may be detected as the language code for the utterances over the conversation. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes.
266
+
"A String",
267
+
],
268
+
"audioEncoding": "A String", # Audio encoding of the audio content to process.
269
+
"enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information.
270
+
"languageCode": "A String", # The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language.
265
271
"model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then Dialogflow auto-selects a model based on other parameters in the SpeechToTextConfig and Agent settings. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search Leave this field unspecified to use [Agent Speech settings](https://cloud.google.com/dialogflow/cx/docs/concept/agent#settings-speech) for model selection.
272
+
"sampleRateHertz": 42, # Sample rate (in Hertz) of the audio content sent in the query. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics) for more details.
266
273
"speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
267
274
"useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.
Copy file name to clipboardExpand all lines: docs/dyn/dialogflow_v2.projects.suggestions.html
+7Lines changed: 7 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -262,7 +262,14 @@ <h3>Method Details</h3>
262
262
},
263
263
"securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
264
264
"sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
265
+
"alternativeLanguageCodes": [ # Defines the list of other language codes in addition to the one provided by the conversation profile that may be detected as the language code for the utterances over the conversation. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes.
266
+
"A String",
267
+
],
268
+
"audioEncoding": "A String", # Audio encoding of the audio content to process.
269
+
"enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information.
270
+
"languageCode": "A String", # The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language.
265
271
"model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then Dialogflow auto-selects a model based on other parameters in the SpeechToTextConfig and Agent settings. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search Leave this field unspecified to use [Agent Speech settings](https://cloud.google.com/dialogflow/cx/docs/concept/agent#settings-speech) for model selection.
272
+
"sampleRateHertz": 42, # Sample rate (in Hertz) of the audio content sent in the query. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics) for more details.
266
273
"speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
267
274
"useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.
0 commit comments