diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html index 42cd727ed0..1150b87c50 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html @@ -272,6 +272,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -288,6 +296,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -305,6 +321,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -449,6 +473,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -591,6 +623,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -610,6 +650,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -754,6 +802,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -911,6 +967,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1062,6 +1126,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1563,6 +1635,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -1579,6 +1659,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1596,6 +1684,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1740,6 +1836,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1882,6 +1986,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1901,6 +2013,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2045,6 +2165,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2202,6 +2330,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2353,6 +2489,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2779,6 +2923,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2796,6 +2948,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2940,6 +3100,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3082,6 +3250,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3101,6 +3277,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3245,6 +3429,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3402,6 +3594,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3553,6 +3753,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3900,6 +4108,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -3916,6 +4132,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3933,6 +4157,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4077,6 +4309,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4219,6 +4459,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4238,6 +4486,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4382,6 +4638,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4539,6 +4803,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4690,6 +4962,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.html b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.html index 8ac1f47fb5..04308ec9e4 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.html @@ -154,6 +154,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -177,6 +185,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -327,6 +343,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -483,6 +507,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -634,6 +666,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -657,6 +697,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -807,6 +855,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -963,6 +1019,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1183,6 +1247,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -1206,6 +1278,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1356,6 +1436,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1512,6 +1600,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1758,6 +1854,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -1781,6 +1885,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1931,6 +2043,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2087,6 +2207,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2256,6 +2384,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2279,6 +2415,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2429,6 +2573,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2585,6 +2737,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2737,6 +2897,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2760,6 +2928,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2910,6 +3086,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3066,6 +3250,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.pages.html b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.pages.html index 54e4becb6e..593454921b 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.pages.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.pages.html @@ -124,6 +124,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -141,6 +149,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -285,6 +301,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -427,6 +451,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -446,6 +478,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -590,6 +630,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -747,6 +795,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -898,6 +954,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1049,6 +1113,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1066,6 +1138,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1210,6 +1290,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1352,6 +1440,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1371,6 +1467,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1515,6 +1619,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1672,6 +1784,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1823,6 +1943,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2000,6 +2128,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2017,6 +2153,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2161,6 +2305,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2303,6 +2455,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -2322,6 +2482,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2466,6 +2634,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2623,6 +2799,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2774,6 +2958,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2937,6 +3129,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2954,6 +3154,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3098,6 +3306,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3240,6 +3456,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3259,6 +3483,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3403,6 +3635,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3560,6 +3800,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3711,6 +3959,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3879,6 +4135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3896,6 +4160,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4040,6 +4312,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4182,6 +4462,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4201,6 +4489,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4345,6 +4641,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4502,6 +4806,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4653,6 +4965,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4805,6 +5125,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -4822,6 +5150,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4966,6 +5302,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5108,6 +5452,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -5127,6 +5479,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5271,6 +5631,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5428,6 +5796,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5579,6 +5955,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.transitionRouteGroups.html b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.transitionRouteGroups.html index 9ce273c049..b034db9b0a 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.flows.transitionRouteGroups.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.flows.transitionRouteGroups.html @@ -135,6 +135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -297,6 +305,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -485,6 +501,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -659,6 +683,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -838,6 +870,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1001,6 +1041,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.html b/docs/dyn/dialogflow_v3.projects.locations.agents.html index a9f619e5f9..a61ac3fab8 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.html @@ -192,6 +192,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -218,6 +226,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -267,6 +280,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -293,6 +314,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -416,6 +442,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -442,6 +476,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -591,6 +630,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -617,6 +664,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -685,6 +737,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -711,6 +771,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -761,6 +826,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -787,6 +860,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html index ff8af72573..7fc8c4b7e0 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html @@ -275,6 +275,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -291,6 +299,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -308,6 +324,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -452,6 +476,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -594,6 +626,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -613,6 +653,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -757,6 +805,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -914,6 +970,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1065,6 +1129,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1566,6 +1638,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -1582,6 +1662,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1599,6 +1687,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1743,6 +1839,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1885,6 +1989,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1904,6 +2016,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2048,6 +2168,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2205,6 +2333,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2356,6 +2492,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2782,6 +2926,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2799,6 +2951,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2943,6 +3103,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3085,6 +3253,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3104,6 +3280,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3248,6 +3432,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3405,6 +3597,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3556,6 +3756,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3903,6 +4111,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -3919,6 +4135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3936,6 +4160,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4080,6 +4312,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4222,6 +4462,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4241,6 +4489,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4385,6 +4641,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4542,6 +4806,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4693,6 +4965,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html index d43f2177b7..66615552ac 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html @@ -249,6 +249,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -400,6 +408,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -554,6 +570,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -695,6 +719,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -718,6 +750,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -868,6 +908,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1024,6 +1072,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1165,6 +1221,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1182,6 +1246,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1326,6 +1398,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1468,6 +1548,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1487,6 +1575,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1631,6 +1727,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1788,6 +1892,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1939,6 +2051,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2082,6 +2202,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2105,6 +2233,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2255,6 +2391,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2411,6 +2555,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2552,6 +2704,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2569,6 +2729,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2713,6 +2881,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2855,6 +3031,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -2874,6 +3058,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3018,6 +3210,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3175,6 +3375,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3326,6 +3534,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3475,6 +3691,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3683,6 +3907,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3700,6 +3932,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3844,6 +4084,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3986,6 +4234,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4005,6 +4261,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4149,6 +4413,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4306,6 +4578,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4457,6 +4737,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4715,6 +5003,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -4732,6 +5028,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4876,6 +5180,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5018,6 +5330,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -5037,6 +5357,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5181,6 +5509,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5338,6 +5674,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5489,6 +5833,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5758,6 +6110,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -5775,6 +6135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5919,6 +6287,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6061,6 +6437,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -6080,6 +6464,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6224,6 +6616,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6381,6 +6781,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6532,6 +6940,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6790,6 +7206,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -6807,6 +7231,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6951,6 +7383,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7093,6 +7533,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -7112,6 +7560,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7256,6 +7712,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7413,6 +7877,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7564,6 +8036,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7884,6 +8364,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -7901,6 +8389,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8045,6 +8541,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8187,6 +8691,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -8206,6 +8718,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8350,6 +8870,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8507,6 +9035,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8658,6 +9194,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8916,6 +9460,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -8933,6 +9485,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9077,6 +9637,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9219,6 +9787,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -9238,6 +9814,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9382,6 +9966,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9539,6 +10131,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9690,6 +10290,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10019,6 +10627,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -10036,6 +10652,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10180,6 +10804,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10322,6 +10954,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -10341,6 +10981,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10485,6 +11133,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10642,6 +11298,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10793,6 +11457,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11051,6 +11723,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -11068,6 +11748,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11212,6 +11900,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11354,6 +12050,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -11373,6 +12077,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11517,6 +12229,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11674,6 +12394,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11825,6 +12553,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12112,6 +12848,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -12129,6 +12873,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12273,6 +13025,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12415,6 +13175,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -12434,6 +13202,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12578,6 +13354,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12735,6 +13519,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12886,6 +13678,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13144,6 +13944,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -13161,6 +13969,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13305,6 +14121,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13447,6 +14271,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -13466,6 +14298,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13610,6 +14450,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13767,6 +14615,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13918,6 +14774,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14188,6 +15052,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -14205,6 +15077,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14349,6 +15229,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14491,6 +15379,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -14510,6 +15406,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14654,6 +15558,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14811,6 +15723,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14962,6 +15882,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15220,6 +16148,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -15237,6 +16173,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15381,6 +16325,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15523,6 +16475,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -15542,6 +16502,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15686,6 +16654,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15843,6 +16819,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15994,6 +16978,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html index cfbd0239e5..26246d1738 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html @@ -166,6 +166,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -183,6 +191,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -327,6 +343,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -469,6 +493,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -488,6 +520,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -632,6 +672,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -789,6 +837,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -940,6 +996,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1216,6 +1280,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1233,6 +1305,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1377,6 +1457,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1519,6 +1607,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1538,6 +1634,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1682,6 +1786,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1839,6 +1951,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1990,6 +2110,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.transitionRouteGroups.html b/docs/dyn/dialogflow_v3.projects.locations.agents.transitionRouteGroups.html index c23d8aac95..f2216e6431 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.transitionRouteGroups.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.transitionRouteGroups.html @@ -135,6 +135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -297,6 +305,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -485,6 +501,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -659,6 +683,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -838,6 +870,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1001,6 +1041,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3.projects.locations.securitySettings.html b/docs/dyn/dialogflow_v3.projects.locations.securitySettings.html index 4fffaca70e..7e43bb20be 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.securitySettings.html +++ b/docs/dyn/dialogflow_v3.projects.locations.securitySettings.html @@ -114,7 +114,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -130,7 +130,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } x__xgafv: string, V1 error format. @@ -145,7 +145,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -161,7 +161,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } @@ -201,7 +201,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -217,7 +217,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } @@ -244,7 +244,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -260,7 +260,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. }, ], } @@ -293,7 +293,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -309,7 +309,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } updateMask: string, Required. The mask to control which fields get updated. If the mask is not present, all fields will be updated. @@ -325,7 +325,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -341,7 +341,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html index 9150883d1b..f89b9de2ef 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html @@ -272,6 +272,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -288,6 +296,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -305,6 +321,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -447,6 +471,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -587,6 +619,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -606,6 +646,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -748,6 +796,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -903,6 +959,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1052,6 +1116,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1562,6 +1634,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -1578,6 +1658,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1595,6 +1683,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1737,6 +1833,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1877,6 +1981,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1896,6 +2008,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2038,6 +2158,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2193,6 +2321,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2342,6 +2478,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2777,6 +2921,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2794,6 +2946,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2936,6 +3096,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3076,6 +3244,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3095,6 +3271,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3237,6 +3421,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3392,6 +3584,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3541,6 +3741,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3886,6 +4094,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -3902,6 +4118,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3919,6 +4143,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4061,6 +4293,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4201,6 +4441,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4220,6 +4468,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4362,6 +4618,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4517,6 +4781,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4666,6 +4938,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.html index 23f8452c3c..8e5d6862ce 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.html @@ -154,6 +154,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -177,6 +185,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -325,6 +341,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -479,6 +503,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -628,6 +660,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -651,6 +691,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -799,6 +847,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -953,6 +1009,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1171,6 +1235,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -1194,6 +1266,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1342,6 +1422,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1496,6 +1584,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1740,6 +1836,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -1763,6 +1867,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1911,6 +2023,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2065,6 +2185,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2232,6 +2360,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2255,6 +2391,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2403,6 +2547,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2557,6 +2709,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2707,6 +2867,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2730,6 +2898,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2878,6 +3054,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3032,6 +3216,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.pages.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.pages.html index 29988218c3..3ef9c76667 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.pages.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.pages.html @@ -124,6 +124,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -141,6 +149,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -283,6 +299,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -423,6 +447,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -442,6 +474,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -584,6 +624,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -739,6 +787,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -888,6 +944,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1037,6 +1101,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1054,6 +1126,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1196,6 +1276,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1336,6 +1424,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1355,6 +1451,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1497,6 +1601,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1652,6 +1764,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1801,6 +1921,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1976,6 +2104,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1993,6 +2129,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2135,6 +2279,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2275,6 +2427,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -2294,6 +2454,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2436,6 +2604,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2591,6 +2767,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2740,6 +2924,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2901,6 +3093,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2918,6 +3118,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3060,6 +3268,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3200,6 +3416,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3219,6 +3443,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3361,6 +3593,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3516,6 +3756,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3665,6 +3913,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3831,6 +4087,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3848,6 +4112,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3990,6 +4262,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4130,6 +4410,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4149,6 +4437,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4291,6 +4587,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4446,6 +4750,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4595,6 +4907,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4745,6 +5065,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -4762,6 +5090,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4904,6 +5240,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5044,6 +5388,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -5063,6 +5415,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5205,6 +5565,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5360,6 +5728,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5509,6 +5885,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.transitionRouteGroups.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.transitionRouteGroups.html index c48847e894..7378a60e7e 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.transitionRouteGroups.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.flows.transitionRouteGroups.html @@ -135,6 +135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -295,6 +303,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -481,6 +497,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -653,6 +677,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -830,6 +862,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -991,6 +1031,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html index 4c114a50cc..d73303e4a6 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html @@ -192,6 +192,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -218,6 +226,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -267,6 +280,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -293,6 +314,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -416,6 +442,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -442,6 +476,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -591,6 +630,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -617,6 +664,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -685,6 +737,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -711,6 +771,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. @@ -761,6 +826,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "answerFeedbackSettings": { # Settings for answer feedback collection. # Optional. Answer feedback collection settings. "enableAnswerFeedback": True or False, # Optional. If enabled, end users will be able to provide answer feedback to Dialogflow responses. Feature works only if interaction logging is enabled in the Dialogflow agent. @@ -787,6 +860,11 @@

Method Details

}, "locked": True or False, # Indicates whether the agent is locked for changes. If the agent is locked, modifications to the agent will be rejected except for RestoreAgent. "name": "A String", # The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`. + "personalizationSettings": { # Settings for end user personalization. # Optional. Settings for end user personalization. + "defaultEndUserMetadata": { # Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { "age": "$session.params.age" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing. + "a_key": "", # Properties of the object. + }, + }, "securitySettings": "A String", # Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`. "speechToTextSettings": { # Settings related to speech recognition. # Speech recognition related settings. "enableSpeechAdaptation": True or False, # Whether to use speech adaptation for speech recognition. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html index aa0a5bf782..e3246591bf 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html @@ -275,6 +275,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -291,6 +299,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -308,6 +324,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -450,6 +474,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -590,6 +622,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -609,6 +649,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -751,6 +799,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -906,6 +962,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1055,6 +1119,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1565,6 +1637,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -1581,6 +1661,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1598,6 +1686,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1740,6 +1836,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1880,6 +1984,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1899,6 +2011,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2041,6 +2161,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2196,6 +2324,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2345,6 +2481,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2780,6 +2924,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2797,6 +2949,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2939,6 +3099,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3079,6 +3247,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3098,6 +3274,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3240,6 +3424,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3395,6 +3587,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3544,6 +3744,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3889,6 +4097,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "allowAnswerFeedback": True or False, # Indicates whether the Thumbs up/Thumbs down rating controls are need to be shown for the response in the Dialogflow Messenger widget. "currentPage": { # A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the [Page guide](https://cloud.google.com/dialogflow/cx/docs/concept/page). # The current Page. Some, not all fields are filled in this message, including but not limited to `name` and `display_name`. @@ -3905,6 +4121,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3922,6 +4146,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4064,6 +4296,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4204,6 +4444,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4223,6 +4471,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4365,6 +4621,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4520,6 +4784,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4669,6 +4941,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html index c14b5bd30b..c2d6050c0b 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html @@ -249,6 +249,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -398,6 +406,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -550,6 +566,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -689,6 +713,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -712,6 +744,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -860,6 +900,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1014,6 +1062,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1153,6 +1209,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1170,6 +1234,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1312,6 +1384,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1452,6 +1532,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1471,6 +1559,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1613,6 +1709,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1768,6 +1872,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1917,6 +2029,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2058,6 +2178,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the flow. The maximum length is 500 characters. If exceeded, the request is rejected. "displayName": "A String", # Required. The human-readable name of the flow. @@ -2081,6 +2209,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2229,6 +2365,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2383,6 +2527,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2522,6 +2674,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -2539,6 +2699,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2681,6 +2849,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2821,6 +2997,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -2840,6 +3024,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -2982,6 +3174,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3137,6 +3337,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3286,6 +3494,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3433,6 +3649,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3639,6 +3863,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -3656,6 +3888,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3798,6 +4038,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -3938,6 +4186,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -3957,6 +4213,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4099,6 +4363,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4254,6 +4526,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4403,6 +4683,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4659,6 +4947,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -4676,6 +4972,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4818,6 +5122,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -4958,6 +5270,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -4977,6 +5297,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5119,6 +5447,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5274,6 +5610,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5423,6 +5767,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5690,6 +6042,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -5707,6 +6067,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5849,6 +6217,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -5989,6 +6365,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -6008,6 +6392,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6150,6 +6542,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6305,6 +6705,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6454,6 +6862,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6710,6 +7126,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -6727,6 +7151,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -6869,6 +7301,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7009,6 +7449,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -7028,6 +7476,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7170,6 +7626,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7325,6 +7789,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7474,6 +7946,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7792,6 +8272,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -7809,6 +8297,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -7951,6 +8447,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8091,6 +8595,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -8110,6 +8622,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8252,6 +8772,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8407,6 +8935,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8556,6 +9092,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8812,6 +9356,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -8829,6 +9381,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -8971,6 +9531,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9111,6 +9679,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -9130,6 +9706,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9272,6 +9856,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9427,6 +10019,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9576,6 +10176,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -9903,6 +10511,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -9920,6 +10536,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10062,6 +10686,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10202,6 +10834,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -10221,6 +10861,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10363,6 +11011,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10518,6 +11174,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10667,6 +11331,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -10923,6 +11595,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -10940,6 +11620,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11082,6 +11770,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11222,6 +11918,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -11241,6 +11945,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11383,6 +12095,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11538,6 +12258,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11687,6 +12415,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -11972,6 +12708,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -11989,6 +12733,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12131,6 +12883,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12271,6 +13031,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -12290,6 +13058,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12432,6 +13208,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12587,6 +13371,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12736,6 +13528,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -12992,6 +13792,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -13009,6 +13817,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13151,6 +13967,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13291,6 +14115,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -13310,6 +14142,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13452,6 +14292,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13607,6 +14455,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -13756,6 +14612,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14024,6 +14888,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -14041,6 +14913,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14183,6 +15063,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14323,6 +15211,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -14342,6 +15238,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14484,6 +15388,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14639,6 +15551,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -14788,6 +15708,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15044,6 +15972,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -15061,6 +15997,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15203,6 +16147,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15343,6 +16295,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -15362,6 +16322,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15504,6 +16472,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15659,6 +16635,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -15808,6 +16792,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html index aba1fc21e7..149c334bf0 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html @@ -166,6 +166,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -183,6 +191,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -325,6 +341,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -465,6 +489,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -484,6 +516,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -626,6 +666,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -781,6 +829,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -930,6 +986,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1204,6 +1268,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "description": "A String", # The description of the page. The maximum length is 500 characters. "displayName": "A String", # Required. The human-readable name of the page, unique within the flow. @@ -1221,6 +1293,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1363,6 +1443,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1503,6 +1591,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "defaultValue": "", # The default value of an optional parameter. If the parameter is required, the default value will be ignored. "displayName": "A String", # Required. The human-readable name of the parameter, unique within the form. @@ -1522,6 +1618,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1664,6 +1768,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1819,6 +1931,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -1968,6 +2088,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.transitionRouteGroups.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.transitionRouteGroups.html index 1b5baa9445..53e2f131f5 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.transitionRouteGroups.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.transitionRouteGroups.html @@ -135,6 +135,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -295,6 +303,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -481,6 +497,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -653,6 +677,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -830,6 +862,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. @@ -991,6 +1031,14 @@

Method Details

"enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. "enableStackdriverLogging": True or False, # If true, StackDriver logging is currently enabled. }, + "speechSettings": { # Define behaviors of speech to text detection. # Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level + "endpointerSensitivity": 42, # Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100. + "models": { # Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models). + "a_key": "A String", + }, + "noSpeechTimeout": "A String", # Timeout before detecting no speech. + "useTimeoutBasedEndpointing": True or False, # Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value. + }, }, "conditionalCases": [ # Conditional cases for this fulfillment. { # A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.securitySettings.html b/docs/dyn/dialogflow_v3beta1.projects.locations.securitySettings.html index 80822adb3c..21fd9bbcc8 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.securitySettings.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.securitySettings.html @@ -114,7 +114,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -130,7 +130,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } x__xgafv: string, V1 error format. @@ -145,7 +145,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -161,7 +161,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } @@ -201,7 +201,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -217,7 +217,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } @@ -244,7 +244,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -260,7 +260,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. }, ], } @@ -293,7 +293,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -309,7 +309,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } updateMask: string, Required. The mask to control which fields get updated. If the mask is not present, all fields will be updated. @@ -325,7 +325,7 @@

Method Details

"audioExportSettings": { # Settings for exporting audio. # Controls audio export settings for post-conversation analytics when ingesting audio to conversations via Participants.AnalyzeContent or Participants.StreamingAnalyzeContent. If retention_strategy is set to REMOVE_AFTER_CONVERSATION or audio_export_settings.gcs_bucket is empty, audio export is disabled. If audio export is enabled, audio is recorded and saved to audio_export_settings.gcs_bucket, subject to retention policy of audio_export_settings.gcs_bucket. This setting won't effect audio input for implicit sessions via Sessions.DetectIntent or Sessions.StreamingDetectIntent. "audioExportPattern": "A String", # Filename pattern for exported audio. "audioFormat": "A String", # File format for exported audio file. Currently only in telephony recordings. - "enableAudioRedaction": True or False, # Enable audio redaction if it is true. + "enableAudioRedaction": True or False, # Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted. "gcsBucket": "A String", # Cloud Storage bucket to export audio record to. Setting this field would grant the Storage Object Creator role to the Dialogflow Service Agent. API caller that tries to modify this field should have the permission of storage.buckets.setIamPolicy. }, "deidentifyTemplate": "A String", # [DLP](https://cloud.google.com/dlp/docs) deidentify template name. Use this template to define de-identification configuration for the content. The `DLP De-identify Templates Reader` role is needed on the Dialogflow service identity service account (has the form `service-PROJECT_NUMBER@gcp-sa-dialogflow.iam.gserviceaccount.com`) for your agent's project. If empty, Dialogflow replaces sensitive info with `[redacted]` text. The template name will have one of the following formats: `projects//locations//deidentifyTemplates/` OR `organizations//locations//deidentifyTemplates/` Note: `deidentify_template` must be located in the same region as the `SecuritySettings`. @@ -341,7 +341,7 @@

Method Details

"redactionScope": "A String", # Defines the data for which Dialogflow applies redaction. Dialogflow does not redact data that it does not have access to – for example, Cloud logging. "redactionStrategy": "A String", # Strategy that defines how we do redaction. "retentionStrategy": "A String", # Specifies the retention behavior defined by SecuritySettings.RetentionStrategy. - "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. + "retentionWindowDays": 42, # Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact. } diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2.json b/googleapiclient/discovery_cache/documents/dialogflow.v2.json index 6436042d5d..2d4e2289b9 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2.json @@ -8327,7 +8327,7 @@ } } }, -"revision": "20240227", +"revision": "20240305", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -8345,6 +8345,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -8384,6 +8388,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3AudioInput", @@ -10623,6 +10655,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -10662,6 +10698,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3beta1AudioInput", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json index 9ccff38bba..8152ff5ff2 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json @@ -7695,7 +7695,7 @@ } } }, -"revision": "20240227", +"revision": "20240305", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -7713,6 +7713,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -7752,6 +7756,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3AudioInput", @@ -9991,6 +10023,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -10030,6 +10066,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3beta1AudioInput", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3.json b/googleapiclient/discovery_cache/documents/dialogflow.v3.json index 22b2ad675c..26e2f7a8fe 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3.json @@ -4453,7 +4453,7 @@ } } }, -"revision": "20240227", +"revision": "20240305", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -4471,6 +4471,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -4510,6 +4514,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3Agent": { "description": "Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, TransitionRouteGroups and so on to manage the conversation flows.", "id": "GoogleCloudDialogflowCxV3Agent", @@ -4563,6 +4595,10 @@ "description": "The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`.", "type": "string" }, +"personalizationSettings": { +"$ref": "GoogleCloudDialogflowCxV3AgentPersonalizationSettings", +"description": "Optional. Settings for end user personalization." +}, "securitySettings": { "description": "Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.", "type": "string" @@ -4656,6 +4692,21 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3AgentPersonalizationSettings": { +"description": "Settings for end user personalization.", +"id": "GoogleCloudDialogflowCxV3AgentPersonalizationSettings", +"properties": { +"defaultEndUserMetadata": { +"additionalProperties": { +"description": "Properties of the object.", +"type": "any" +}, +"description": "Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { \"age\": \"$session.params.age\" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing.", +"type": "object" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3AgentValidationResult": { "description": "The response message for Agents.GetAgentValidationResult.", "id": "GoogleCloudDialogflowCxV3AgentValidationResult", @@ -8593,7 +8644,7 @@ false "type": "string" }, "retentionWindowDays": { -"description": "Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL.", +"description": "Retains the data for the specified number of days. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact.", "format": "int32", "type": "integer" } @@ -8625,7 +8676,7 @@ false "type": "string" }, "enableAudioRedaction": { -"description": "Enable audio redaction if it is true.", +"description": "Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted.", "type": "boolean" }, "gcsBucket": { @@ -9829,6 +9880,10 @@ false "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -9868,6 +9923,34 @@ false }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3beta1AudioInput", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json index d630ad65dc..d81cf4bc11 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json @@ -4453,7 +4453,7 @@ } } }, -"revision": "20240227", +"revision": "20240305", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -4471,6 +4471,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -4510,6 +4514,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3AudioInput": { "description": "Represents the natural speech audio to be processed.", "id": "GoogleCloudDialogflowCxV3AudioInput", @@ -6749,6 +6781,10 @@ "loggingSettings": { "$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsLoggingSettings", "description": "Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level." +}, +"speechSettings": { +"$ref": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"description": "Settings for speech to text detection. Exposed at the following levels: - Agent level - Flow level - Page level - Parameter level" } }, "type": "object" @@ -6788,6 +6824,34 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings": { +"description": "Define behaviors of speech to text detection.", +"id": "GoogleCloudDialogflowCxV3beta1AdvancedSettingsSpeechSettings", +"properties": { +"endpointerSensitivity": { +"description": "Sensitivity of the speech model that detects the end of speech. Scale from 0 to 100.", +"format": "int32", +"type": "integer" +}, +"models": { +"additionalProperties": { +"type": "string" +}, +"description": "Mapping from language to Speech-to-Text model. The mapped Speech-to-Text model will be selected for requests from its corresponding language. For more information, see [Speech models](https://cloud.google.com/dialogflow/cx/docs/concept/speech-models).", +"type": "object" +}, +"noSpeechTimeout": { +"description": "Timeout before detecting no speech.", +"format": "google-duration", +"type": "string" +}, +"useTimeoutBasedEndpointing": { +"description": "Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1Agent": { "description": "Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, TransitionRouteGroups and so on to manage the conversation flows.", "id": "GoogleCloudDialogflowCxV3beta1Agent", @@ -6841,6 +6905,10 @@ "description": "The unique identifier of the agent. Required for the Agents.UpdateAgent method. Agents.CreateAgent populates the name automatically. Format: `projects//locations//agents/`.", "type": "string" }, +"personalizationSettings": { +"$ref": "GoogleCloudDialogflowCxV3beta1AgentPersonalizationSettings", +"description": "Optional. Settings for end user personalization." +}, "securitySettings": { "description": "Name of the SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.", "type": "string" @@ -6934,6 +7002,21 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1AgentPersonalizationSettings": { +"description": "Settings for end user personalization.", +"id": "GoogleCloudDialogflowCxV3beta1AgentPersonalizationSettings", +"properties": { +"defaultEndUserMetadata": { +"additionalProperties": { +"description": "Properties of the object.", +"type": "any" +}, +"description": "Optional. Default end user metadata, used when processing DetectIntent requests. Recommended to be filled as a template instead of hard-coded value, for example { \"age\": \"$session.params.age\" }. The data will be merged with the QueryParameters.end_user_metadata in DetectIntentRequest.query_params during query processing.", +"type": "object" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1AgentValidationResult": { "description": "The response message for Agents.GetAgentValidationResult.", "id": "GoogleCloudDialogflowCxV3beta1AgentValidationResult", @@ -10884,7 +10967,7 @@ false "type": "string" }, "retentionWindowDays": { -"description": "Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL.", +"description": "Retains data in interaction logging for the specified number of days. This does not apply to Cloud logging, which is owned by the user - not Dialogflow. User must set a value lower than Dialogflow's default 365d TTL (30 days for Agent Assist traffic), higher value will be ignored and use default. Setting a value higher than that has no effect. A missing value or setting to 0 also means we use default TTL. When data retention configuration is changed, it only applies to the data created after the change; the TTL of existing data created before the change stays intact.", "format": "int32", "type": "integer" } @@ -10916,7 +10999,7 @@ false "type": "string" }, "enableAudioRedaction": { -"description": "Enable audio redaction if it is true.", +"description": "Enable audio redaction if it is true. Note that this only redacts end-user audio data; Synthesised audio from the virtual agent is not redacted.", "type": "boolean" }, "gcsBucket": {