diff --git a/docs/dyn/dialogflow_v2.projects.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2.projects.agent.environments.users.sessions.html index 276750ee497..e8d8b804583 100644 --- a/docs/dyn/dialogflow_v2.projects.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2.projects.agent.environments.users.sessions.html @@ -149,6 +149,7 @@

Method Details

"audioConfig": { # Instructs the speech recognizer how to process the audio content. # Instructs the speech recognizer how to process the speech audio. "audioEncoding": "A String", # Required. Audio encoding of the audio content to process. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2.projects.agent.sessions.html b/docs/dyn/dialogflow_v2.projects.agent.sessions.html index 50cde4b2b64..928972b8903 100644 --- a/docs/dyn/dialogflow_v2.projects.agent.sessions.html +++ b/docs/dyn/dialogflow_v2.projects.agent.sessions.html @@ -149,6 +149,7 @@

Method Details

"audioConfig": { # Instructs the speech recognizer how to process the audio content. # Instructs the speech recognizer how to process the speech audio. "audioEncoding": "A String", # Required. Audio encoding of the audio content to process. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2.projects.answerRecords.html b/docs/dyn/dialogflow_v2.projects.answerRecords.html index 2dfab78b536..59367f3bb8e 100644 --- a/docs/dyn/dialogflow_v2.projects.answerRecords.html +++ b/docs/dyn/dialogflow_v2.projects.answerRecords.html @@ -148,7 +148,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -220,7 +220,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -274,7 +274,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. diff --git a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html index c3cbc939144..aac78f964f6 100644 --- a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html @@ -170,6 +170,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -177,7 +178,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -213,6 +214,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -220,7 +222,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -325,6 +327,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -332,7 +335,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -368,6 +371,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -375,7 +379,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -505,6 +509,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -512,7 +517,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -548,6 +553,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -555,7 +561,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -671,6 +677,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -678,7 +685,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -714,6 +721,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -721,7 +729,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -845,6 +853,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -852,7 +861,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -888,6 +897,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -895,7 +905,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1001,6 +1011,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1008,7 +1019,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1044,6 +1055,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1051,7 +1063,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1151,6 +1163,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1158,7 +1171,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. diff --git a/docs/dyn/dialogflow_v2.projects.conversations.suggestions.html b/docs/dyn/dialogflow_v2.projects.conversations.suggestions.html index 95d36be9bbe..844c3a8f4f0 100644 --- a/docs/dyn/dialogflow_v2.projects.conversations.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.conversations.suggestions.html @@ -96,11 +96,6 @@

Method Details

The object takes the form of: { # The request message for Conversations.SuggestConversationSummary. - "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. - "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` - "a_key": "A String", - }, - }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000. "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. If empty, the latest message of the conversation will be used. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2.projects.html b/docs/dyn/dialogflow_v2.projects.html index 2a88265f222..6cb082c1778 100644 --- a/docs/dyn/dialogflow_v2.projects.html +++ b/docs/dyn/dialogflow_v2.projects.html @@ -119,6 +119,11 @@

Instance Methods

Returns the operations Resource.

+

+ suggestions() +

+

Returns the suggestions Resource.

+

close()

Close httplib2 connections.

diff --git a/docs/dyn/dialogflow_v2.projects.locations.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2.projects.locations.agent.environments.users.sessions.html index 46de77cff80..9132300494f 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.agent.environments.users.sessions.html @@ -149,6 +149,7 @@

Method Details

"audioConfig": { # Instructs the speech recognizer how to process the audio content. # Instructs the speech recognizer how to process the speech audio. "audioEncoding": "A String", # Required. Audio encoding of the audio content to process. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2.projects.locations.agent.sessions.html b/docs/dyn/dialogflow_v2.projects.locations.agent.sessions.html index 9e62db3544f..5e2da3a4d2c 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.agent.sessions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.agent.sessions.html @@ -149,6 +149,7 @@

Method Details

"audioConfig": { # Instructs the speech recognizer how to process the audio content. # Instructs the speech recognizer how to process the speech audio. "audioEncoding": "A String", # Required. Audio encoding of the audio content to process. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2.projects.locations.answerRecords.html b/docs/dyn/dialogflow_v2.projects.locations.answerRecords.html index 8e26af5860c..043a38f342c 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.answerRecords.html +++ b/docs/dyn/dialogflow_v2.projects.locations.answerRecords.html @@ -148,7 +148,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -220,7 +220,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -274,7 +274,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html index 51ddbf78f8c..760fa13ffb4 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html @@ -170,6 +170,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -177,7 +178,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -213,6 +214,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -220,7 +222,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -325,6 +327,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -332,7 +335,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -368,6 +371,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -375,7 +379,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -505,6 +509,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -512,7 +517,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -548,6 +553,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -555,7 +561,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -671,6 +677,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -678,7 +685,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -714,6 +721,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -721,7 +729,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -845,6 +853,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -852,7 +861,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -888,6 +897,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -895,7 +905,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1001,6 +1011,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1008,7 +1019,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1044,6 +1055,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1051,7 +1063,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1151,6 +1163,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1158,7 +1171,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversations.suggestions.html b/docs/dyn/dialogflow_v2.projects.locations.conversations.suggestions.html index b856d1503ac..d0220f8208e 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversations.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversations.suggestions.html @@ -96,11 +96,6 @@

Method Details

The object takes the form of: { # The request message for Conversations.SuggestConversationSummary. - "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. - "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` - "a_key": "A String", - }, - }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000. "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. If empty, the latest message of the conversation will be used. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2.projects.locations.html b/docs/dyn/dialogflow_v2.projects.locations.html index 0a82cc8b691..98b78dd3c20 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.html +++ b/docs/dyn/dialogflow_v2.projects.locations.html @@ -114,6 +114,11 @@

Instance Methods

Returns the operations Resource.

+

+ suggestions() +

+

Returns the suggestions Resource.

+

close()

Close httplib2 connections.

@@ -173,7 +178,7 @@

Method Details

Returns: An object of the form: - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", @@ -237,7 +242,7 @@

Method Details

{ # The response message for Locations.ListLocations. "locations": [ # A list of locations that matches the specified filter in the request. - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", diff --git a/docs/dyn/dialogflow_v2.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html new file mode 100644 index 00000000000..a9c361f3a48 --- /dev/null +++ b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html @@ -0,0 +1,300 @@ + + + +

Dialogflow API . projects . locations . suggestions

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ generateStatelessSummary(parent, body=None, x__xgafv=None)

+

Generates and returns a summary for a conversation that does not have a resource created for it.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ generateStatelessSummary(parent, body=None, x__xgafv=None) +
Generates and returns a summary for a conversation that does not have a resource created for it.
+
+Args:
+  parent: string, Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # The request message for Conversations.GenerateStatelessSummary.
+  "conversationProfile": { # Defines the services to connect to incoming Dialogflow conversations. # Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}
+    "automatedAgentConfig": { # Defines the Automated Agent to connect to a conversation. # Configuration for an automated agent to use with this profile.
+      "agent": "A String", # Required. ID of the Dialogflow agent environment to use. This project needs to either be the same project as the conversation or you need to grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow API Service Agent` role in this project. - For ES agents, use format: `projects//locations//agent/environments/`. If environment is not specified, the default `draft` environment is used. Refer to [DetectIntentRequest](/dialogflow/docs/reference/rpc/google.cloud.dialogflow.v2#google.cloud.dialogflow.v2.DetectIntentRequest) for more details. - For CX agents, use format `projects//locations//agents//environments/`. If environment is not specified, the default `draft` environment is used.
+    },
+    "createTime": "A String", # Output only. Create time of the conversation profile.
+    "displayName": "A String", # Required. Human readable name for this profile. Max length 1024 bytes.
+    "humanAgentAssistantConfig": { # Defines the Human Agent Assist to connect to a conversation. # Configuration for agent assistance to use with this profile.
+      "endUserSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of end user participant. Currently, this feature is not general available, please contact Google to get access.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a Dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, at most 5 documents are supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, at most 5 knowledge bases are supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION and FAQ will use this field.
+              "noSmalltalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a Dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, at most 5 documents are supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, at most 5 knowledge bases are supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION and FAQ will use this field.
+              "noSmalltalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis.
+        "enableEntityExtraction": True or False, # Enable entity extraction in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Currently, this feature is not general available, please contact Google to get access.
+        "enableSentimentAnalysis": True or False, # Enable sentiment analysis in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral: https://cloud.google.com/natural-language/docs/basics#sentiment_analysis For Participants.StreamingAnalyzeContent method, result will be in StreamingAnalyzeContentResponse.message.SentimentAnalysisResult. For Participants.AnalyzeContent method, result will be in AnalyzeContentResponse.message.SentimentAnalysisResult For Conversations.ListMessages method, result will be in ListMessagesResponse.messages.SentimentAnalysisResult If Pub/Sub notification is configured, result will be in ConversationEvent.new_message_payload.SentimentAnalysisResult.
+      },
+      "notificationConfig": { # Defines notification behavior. # Pub/Sub topic on which to publish new agent assistant events.
+        "messageFormat": "A String", # Format of message.
+        "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+      },
+    },
+    "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access.
+      "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com).
+        "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page.
+      },
+      "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent.
+        "buttonId": "A String", # Required. Live Agent chat button ID.
+        "deploymentId": "A String", # Required. Live Agent deployment ID.
+        "endpointDomain": "A String", # Required. Domain of the Live Agent endpoint for this agent. You can find the endpoint URL in the `Live Agent settings` page. For example if URL has the form https://d.la4-c2-phx.salesforceliveagent.com/..., you should fill in d.la4-c2-phx.salesforceliveagent.com.
+        "organizationId": "A String", # Required. The organization ID of the Salesforce account.
+      },
+    },
+    "languageCode": "A String", # Language code for the conversation profile. If not specified, the language is en-US. Language at ConversationProfile should be set for all non en-US languages. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+    "loggingConfig": { # Defines logging behavior for conversation lifecycle events. # Configuration for logging conversation lifecycle events.
+      "enableStackdriverLogging": True or False, # Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.
+    },
+    "name": "A String", # The unique identifier of this conversation profile. Format: `projects//locations//conversationProfiles/`.
+    "newMessageEventNotificationConfig": { # Defines notification behavior. # Configuration for publishing new message events. Event will be sent in format of ConversationEvent
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "notificationConfig": { # Defines notification behavior. # Configuration for publishing conversation lifecycle events.
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
+    "sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
+      "model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then a default model is used. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details.
+      "speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
+    },
+    "timeZone": "A String", # The time zone of this conversational profile from the [time zone database](https://www.iana.org/time-zones), e.g., America/New_York, Europe/Paris. Defaults to America/New_York.
+    "ttsConfig": { # Configuration of how speech should be synthesized. # Configuration for Text-to-Speech synthesization. Used by Phone Gateway to specify synthesization options. If agent defines synthesization options as well, agent settings overrides the option here.
+      "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
+        "A String",
+      ],
+      "pitch": 3.14, # Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
+      "speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
+      "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio.
+        "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender.
+        "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
+      },
+      "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
+    },
+    "updateTime": "A String", # Output only. Update time of the conversation profile.
+  },
+  "latestMessage": "A String", # The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.
+  "maxContextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.
+  "statelessConversation": { # The minimum amount of information required to generate a Summary without having a Conversation resource created. # Required. The conversation to suggest a summary for.
+    "messages": [ # Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.
+      { # Represents a message posted into a conversation.
+        "content": "A String", # Required. The message content.
+        "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
+        "languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+        "messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
+          "containEntities": True or False, # Indicates whether the text message contains entities.
+          "parts": [ # The collection of annotated message parts ordered by their position in the message. You can recover the annotated message by concatenating [AnnotatedMessagePart.text].
+            { # Represents a part of a message possibly annotated with an entity. The part can be an entity or purely a part of the message between two entities or message start/end.
+              "entityType": "A String", # The [Dialogflow system entity type](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. If this is empty, Dialogflow could not annotate the phrase part with a system entity.
+              "formattedValue": "", # The [Dialogflow system entity formatted value ](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. For example for a system entity of type `@sys.unit-currency`, this may contain: { "amount": 5, "currency": "USD" }
+              "text": "A String", # A part of a message possibly annotated with an entity.
+            },
+          ],
+        },
+        "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+        "participant": "A String", # Output only. The participant that sends this message.
+        "participantRole": "A String", # Output only. The role of the participant.
+        "sendTime": "A String", # Optional. The time when the message was sent.
+        "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+          "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`.
+            "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+            "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+          },
+        },
+      },
+    ],
+  },
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # The response message for Conversations.GenerateStatelessSummary.
+  "contextSize": 42, # Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.
+  "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.
+  "summary": { # Generated summary for a conversation. # Generated summary.
+    "text": "A String", # The summary content that is concatenated into one string.
+    "textSections": { # The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.
+      "a_key": "A String",
+    },
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2.projects.suggestions.html b/docs/dyn/dialogflow_v2.projects.suggestions.html new file mode 100644 index 00000000000..178df639d28 --- /dev/null +++ b/docs/dyn/dialogflow_v2.projects.suggestions.html @@ -0,0 +1,300 @@ + + + +

Dialogflow API . projects . suggestions

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ generateStatelessSummary(parent, body=None, x__xgafv=None)

+

Generates and returns a summary for a conversation that does not have a resource created for it.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ generateStatelessSummary(parent, body=None, x__xgafv=None) +
Generates and returns a summary for a conversation that does not have a resource created for it.
+
+Args:
+  parent: string, Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # The request message for Conversations.GenerateStatelessSummary.
+  "conversationProfile": { # Defines the services to connect to incoming Dialogflow conversations. # Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}
+    "automatedAgentConfig": { # Defines the Automated Agent to connect to a conversation. # Configuration for an automated agent to use with this profile.
+      "agent": "A String", # Required. ID of the Dialogflow agent environment to use. This project needs to either be the same project as the conversation or you need to grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow API Service Agent` role in this project. - For ES agents, use format: `projects//locations//agent/environments/`. If environment is not specified, the default `draft` environment is used. Refer to [DetectIntentRequest](/dialogflow/docs/reference/rpc/google.cloud.dialogflow.v2#google.cloud.dialogflow.v2.DetectIntentRequest) for more details. - For CX agents, use format `projects//locations//agents//environments/`. If environment is not specified, the default `draft` environment is used.
+    },
+    "createTime": "A String", # Output only. Create time of the conversation profile.
+    "displayName": "A String", # Required. Human readable name for this profile. Max length 1024 bytes.
+    "humanAgentAssistantConfig": { # Defines the Human Agent Assist to connect to a conversation. # Configuration for agent assistance to use with this profile.
+      "endUserSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of end user participant. Currently, this feature is not general available, please contact Google to get access.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a Dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, at most 5 documents are supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, at most 5 knowledge bases are supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION and FAQ will use this field.
+              "noSmalltalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a Dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, at most 5 documents are supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, at most 5 knowledge bases are supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION and FAQ will use this field.
+              "noSmalltalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis.
+        "enableEntityExtraction": True or False, # Enable entity extraction in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Currently, this feature is not general available, please contact Google to get access.
+        "enableSentimentAnalysis": True or False, # Enable sentiment analysis in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral: https://cloud.google.com/natural-language/docs/basics#sentiment_analysis For Participants.StreamingAnalyzeContent method, result will be in StreamingAnalyzeContentResponse.message.SentimentAnalysisResult. For Participants.AnalyzeContent method, result will be in AnalyzeContentResponse.message.SentimentAnalysisResult For Conversations.ListMessages method, result will be in ListMessagesResponse.messages.SentimentAnalysisResult If Pub/Sub notification is configured, result will be in ConversationEvent.new_message_payload.SentimentAnalysisResult.
+      },
+      "notificationConfig": { # Defines notification behavior. # Pub/Sub topic on which to publish new agent assistant events.
+        "messageFormat": "A String", # Format of message.
+        "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+      },
+    },
+    "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access.
+      "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com).
+        "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page.
+      },
+      "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent.
+        "buttonId": "A String", # Required. Live Agent chat button ID.
+        "deploymentId": "A String", # Required. Live Agent deployment ID.
+        "endpointDomain": "A String", # Required. Domain of the Live Agent endpoint for this agent. You can find the endpoint URL in the `Live Agent settings` page. For example if URL has the form https://d.la4-c2-phx.salesforceliveagent.com/..., you should fill in d.la4-c2-phx.salesforceliveagent.com.
+        "organizationId": "A String", # Required. The organization ID of the Salesforce account.
+      },
+    },
+    "languageCode": "A String", # Language code for the conversation profile. If not specified, the language is en-US. Language at ConversationProfile should be set for all non en-US languages. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+    "loggingConfig": { # Defines logging behavior for conversation lifecycle events. # Configuration for logging conversation lifecycle events.
+      "enableStackdriverLogging": True or False, # Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.
+    },
+    "name": "A String", # The unique identifier of this conversation profile. Format: `projects//locations//conversationProfiles/`.
+    "newMessageEventNotificationConfig": { # Defines notification behavior. # Configuration for publishing new message events. Event will be sent in format of ConversationEvent
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "notificationConfig": { # Defines notification behavior. # Configuration for publishing conversation lifecycle events.
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
+    "sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
+      "model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then a default model is used. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details.
+      "speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
+    },
+    "timeZone": "A String", # The time zone of this conversational profile from the [time zone database](https://www.iana.org/time-zones), e.g., America/New_York, Europe/Paris. Defaults to America/New_York.
+    "ttsConfig": { # Configuration of how speech should be synthesized. # Configuration for Text-to-Speech synthesization. Used by Phone Gateway to specify synthesization options. If agent defines synthesization options as well, agent settings overrides the option here.
+      "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
+        "A String",
+      ],
+      "pitch": 3.14, # Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
+      "speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
+      "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio.
+        "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender.
+        "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
+      },
+      "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
+    },
+    "updateTime": "A String", # Output only. Update time of the conversation profile.
+  },
+  "latestMessage": "A String", # The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.
+  "maxContextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.
+  "statelessConversation": { # The minimum amount of information required to generate a Summary without having a Conversation resource created. # Required. The conversation to suggest a summary for.
+    "messages": [ # Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.
+      { # Represents a message posted into a conversation.
+        "content": "A String", # Required. The message content.
+        "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
+        "languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+        "messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
+          "containEntities": True or False, # Indicates whether the text message contains entities.
+          "parts": [ # The collection of annotated message parts ordered by their position in the message. You can recover the annotated message by concatenating [AnnotatedMessagePart.text].
+            { # Represents a part of a message possibly annotated with an entity. The part can be an entity or purely a part of the message between two entities or message start/end.
+              "entityType": "A String", # The [Dialogflow system entity type](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. If this is empty, Dialogflow could not annotate the phrase part with a system entity.
+              "formattedValue": "", # The [Dialogflow system entity formatted value ](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. For example for a system entity of type `@sys.unit-currency`, this may contain: { "amount": 5, "currency": "USD" }
+              "text": "A String", # A part of a message possibly annotated with an entity.
+            },
+          ],
+        },
+        "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+        "participant": "A String", # Output only. The participant that sends this message.
+        "participantRole": "A String", # Output only. The role of the participant.
+        "sendTime": "A String", # Optional. The time when the message was sent.
+        "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+          "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`.
+            "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+            "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+          },
+        },
+      },
+    ],
+  },
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # The response message for Conversations.GenerateStatelessSummary.
+  "contextSize": 42, # Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.
+  "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.
+  "summary": { # Generated summary for a conversation. # Generated summary.
+    "text": "A String", # The summary content that is concatenated into one string.
+    "textSections": { # The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.
+      "a_key": "A String",
+    },
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html index fb431aad382..4e0b6f202d3 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html @@ -153,6 +153,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html index 14f3a4c30a7..edd6592922d 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html @@ -153,6 +153,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2beta1.projects.answerRecords.html b/docs/dyn/dialogflow_v2beta1.projects.answerRecords.html index 9d85729b6cb..67003c9a629 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.answerRecords.html +++ b/docs/dyn/dialogflow_v2beta1.projects.answerRecords.html @@ -122,6 +122,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -145,7 +821,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -186,6 +862,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -209,7 +1561,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -257,6 +1609,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -280,7 +2308,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -310,6 +2338,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -333,7 +3037,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html index f73ec3a217e..32341955898 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html @@ -170,6 +170,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -177,7 +178,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -185,6 +186,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -213,6 +217,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -220,7 +225,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -228,6 +233,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -325,6 +333,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -332,7 +341,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -340,6 +349,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -368,6 +380,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -375,7 +388,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -383,6 +396,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -505,6 +521,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -512,7 +529,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -520,6 +537,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -548,6 +568,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -555,7 +576,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -563,6 +584,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -671,6 +695,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -678,7 +703,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -686,6 +711,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -714,6 +742,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -721,7 +750,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -729,6 +758,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -845,6 +877,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -852,7 +885,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -860,6 +893,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -888,6 +924,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -895,7 +932,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -903,6 +940,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1001,6 +1041,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1008,7 +1049,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1016,6 +1057,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1044,6 +1088,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1051,7 +1096,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1059,6 +1104,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1151,6 +1199,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1158,7 +1207,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1166,6 +1215,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html index 7d976504e20..91d1fb291bb 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html @@ -125,6 +125,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) @@ -222,6 +223,20 @@

Method Details

}, }, "requestId": "A String", # A unique identifier for this request. Restricted to 36 ASCII characters. A random UUID is recommended. This request is only idempotent if a `request_id` is provided. + "suggestionInput": { # Represents the selection of a suggestion. # An input representing the selection of a suggestion. + "answerRecord": "A String", # Required. The ID of a suggestion selected by the human agent. The suggestion(s) were generated in a previous call to request Dialogflow assist. The format is: `projects//locations//answerRecords/` where is an alphanumeric string. + "intentInput": { # Represents the intent to trigger programmatically rather than as a result of natural language processing. The intent input is only used for V3 agent. # The intent to be triggered on V3 agent. + "intent": "A String", # Required. The unique identifier of the intent in V3 agent. Format: `projects//locations//locations//agents//intents/`. + "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + }, + "parameters": { # In Dialogflow assist for v3, the user can submit a form by sending a SuggestionInput. The form is uniquely determined by the answer_record field, which identifies a v3 QueryResult containing the current page. The form parameters are specified via the parameters field. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "textOverride": { # Represents the natural language text to be processed. # Optional. If the customer edited the suggestion before using it, include the revised text here. + "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. + "text": "A String", # Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters for virtual agent interactions. + }, + }, "textInput": { # Represents the natural language text to be processed. # The natural language text to be processed. "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "text": "A String", # Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters for virtual agent interactions. @@ -1681,62 +1696,1426 @@

Method Details

"contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. }, - "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. - "faqAnswers": [ # Output only. Answers extracted from FAQ documents. - { # Represents answer from "frequently asked questions". - "answer": "A String", # The piece of text from the `source` knowledge base document. - "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "confidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query, range from 0.0 (completely uncertain) to 1.0 (completely certain). - "metadata": { # A map that contains metadata about the answer and the document from which it originates. - "a_key": "A String", - }, - "question": "A String", # The corresponding FAQ question. - "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//locations//agent/knowledgeBases//documents/`. - }, - ], - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. - }, - "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY. - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation. - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. - "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig. - { # Represents a smart reply answer. - "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain). - "reply": "A String", # The content of the reply. - }, - ], - }, - }, - ], - "humanAgentSuggestionResults": [ # The suggestions for most recent human agent. The order is the same as HumanAgentAssistantConfig.SuggestionConfig.feature_configs of HumanAgentAssistantConfig.human_agent_suggestion_config. Note that any failure of Agent Assist features will not lead to the overall failure of an AnalyzeContent API call. Instead, the features will fail silently with the error field set in the corresponding SuggestionResult. - { # One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent. - "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Error status if the request failed. - "code": 42, # The status code, which should be an enum value of google.rpc.Code. - "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. - { - "a_key": "", # Properties of the object. Contains field @type with type URL. - }, - ], - "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. - }, - "suggestArticlesResponse": { # The response message for Participants.SuggestArticles. # SuggestArticlesResponse if request is for ARTICLE_SUGGESTION. - "articleAnswers": [ # Output only. Articles ordered by score in descending order. - { # Represents article answer. + "suggestDialogflowAssistsResponse": { # The response message for Participants.SuggestDialogflowAssists. # SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation. + "dialogflowAssistAnswers": [ # Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction. + { # Represents a Dialogflow assist answer. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "metadata": { # A map that contains metadata about the answer and the document from which it originates. - "a_key": "A String", + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. }, - "snippets": [ # Output only. Article snippets. - "A String", - ], - "title": "A String", # The article title. - "uri": "A String", # The article URI. - }, - ], - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`. + }, + "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. + "faqAnswers": [ # Output only. Answers extracted from FAQ documents. + { # Represents answer from "frequently asked questions". + "answer": "A String", # The piece of text from the `source` knowledge base document. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "confidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query, range from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # A map that contains metadata about the answer and the document from which it originates. + "a_key": "A String", + }, + "question": "A String", # The corresponding FAQ question. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//locations//agent/knowledgeBases//documents/`. + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + }, + "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation. + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig. + { # Represents a smart reply answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain). + "reply": "A String", # The content of the reply. + }, + ], + }, + }, + ], + "humanAgentSuggestionResults": [ # The suggestions for most recent human agent. The order is the same as HumanAgentAssistantConfig.SuggestionConfig.feature_configs of HumanAgentAssistantConfig.human_agent_suggestion_config. Note that any failure of Agent Assist features will not lead to the overall failure of an AnalyzeContent API call. Instead, the features will fail silently with the error field set in the corresponding SuggestionResult. + { # One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Error status if the request failed. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "suggestArticlesResponse": { # The response message for Participants.SuggestArticles. # SuggestArticlesResponse if request is for ARTICLE_SUGGESTION. + "articleAnswers": [ # Output only. Articles ordered by score in descending order. + { # Represents article answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "metadata": { # A map that contains metadata about the answer and the document from which it originates. + "a_key": "A String", + }, + "snippets": [ # Output only. Article snippets. + "A String", + ], + "title": "A String", # The article title. + "uri": "A String", # The article URI. + }, + ], + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + }, + "suggestDialogflowAssistsResponse": { # The response message for Participants.SuggestDialogflowAssists. # SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation. + "dialogflowAssistAnswers": [ # Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction. + { # Represents a Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`. }, "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.suggestions.html index 5896b3ac216..697495dedeb 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.suggestions.html @@ -96,11 +96,6 @@

Method Details

The object takes the form of: { # The request message for Conversations.SuggestConversationSummary. - "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. - "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` - "a_key": "A String", - }, - }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000. "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. If empty, the latest message of the conversation will be used. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2beta1.projects.html b/docs/dyn/dialogflow_v2beta1.projects.html index 504d778bafd..13f3be6396e 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.html +++ b/docs/dyn/dialogflow_v2beta1.projects.html @@ -109,6 +109,11 @@

Instance Methods

Returns the operations Resource.

+

+ suggestions() +

+

Returns the suggestions Resource.

+

close()

Close httplib2 connections.

diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html index f329d4a683a..a9b4de0343c 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html @@ -153,6 +153,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html index 5165f78f7d1..bc07a3cbda8 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html @@ -153,6 +153,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.answerRecords.html b/docs/dyn/dialogflow_v2beta1.projects.locations.answerRecords.html index 943a05d75e2..f49942368f0 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.answerRecords.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.answerRecords.html @@ -122,6 +122,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -145,7 +821,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -186,6 +862,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -209,7 +1561,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -257,6 +1609,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -280,7 +2308,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. @@ -310,6 +2338,682 @@

Method Details

"title": "A String", # The article title. "uri": "A String", # The article URI. }, + "dialogflowAssistAnswer": { # Represents a Dialogflow assist answer. # Output only. The Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, "faqAnswer": { # Represents answer from "frequently asked questions". # Output only. The FAQ answer. "answer": "A String", # The piece of text from the `source` knowledge base document. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" @@ -333,7 +3037,7 @@

Method Details

}, }, "clickTime": "A String", # Time when the answer/item was clicked. - "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. + "clicked": True or False, # Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked. "correctnessLevel": "A String", # The correctness level of the specific answer. "displayTime": "A String", # Time when the answer/item was displayed. "displayed": True or False, # Indicates whether the answer/item was displayed to the human agent in the agent desktop UI. Default to false. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html index 54dc7b6b43f..33f9e73993f 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html @@ -170,6 +170,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -177,7 +178,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -185,6 +186,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -213,6 +217,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -220,7 +225,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -228,6 +233,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -325,6 +333,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -332,7 +341,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -340,6 +349,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -368,6 +380,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -375,7 +388,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -383,6 +396,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -505,6 +521,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -512,7 +529,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -520,6 +537,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -548,6 +568,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -555,7 +576,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -563,6 +584,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -671,6 +695,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -678,7 +703,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -686,6 +711,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -714,6 +742,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -721,7 +750,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -729,6 +758,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -845,6 +877,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -852,7 +885,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -860,6 +893,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -888,6 +924,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -895,7 +932,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -903,6 +940,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1001,6 +1041,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1008,7 +1049,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1016,6 +1057,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1044,6 +1088,7 @@

Method Details

"featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config. { # Config for suggestion features. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1051,7 +1096,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1059,6 +1104,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. @@ -1151,6 +1199,7 @@

Method Details

"participantRole": "A String", # Required. The participant role to add or update the suggestion feature config. Only HUMAN_AGENT or END_USER can be used. "suggestionFeatureConfig": { # Config for suggestion features. # Required. The suggestion feature config to add or update. "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model. + "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0 "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`. }, "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation. @@ -1158,7 +1207,7 @@

Method Details

}, "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST. "queryConfig": { # Config for suggestion query. # Configs of query. - "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE. + "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION. "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped. "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped. "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped. @@ -1166,6 +1215,9 @@

Method Details

}, "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST. "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project. + "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent. + "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`. + }, }, "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE. "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html index e10f3882fa0..f01f58437ee 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html @@ -125,6 +125,7 @@

Method Details

"totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. + "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. "languageCode": "A String", # Required. The language of the supplied audio. Dialogflow does not do translations. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "model": "A String", # Which Speech model to select for the given request. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then we auto-select a model based on the parameters in the InputAudioConfig. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search (best for very short utterances and commands) @@ -222,6 +223,20 @@

Method Details

}, }, "requestId": "A String", # A unique identifier for this request. Restricted to 36 ASCII characters. A random UUID is recommended. This request is only idempotent if a `request_id` is provided. + "suggestionInput": { # Represents the selection of a suggestion. # An input representing the selection of a suggestion. + "answerRecord": "A String", # Required. The ID of a suggestion selected by the human agent. The suggestion(s) were generated in a previous call to request Dialogflow assist. The format is: `projects//locations//answerRecords/` where is an alphanumeric string. + "intentInput": { # Represents the intent to trigger programmatically rather than as a result of natural language processing. The intent input is only used for V3 agent. # The intent to be triggered on V3 agent. + "intent": "A String", # Required. The unique identifier of the intent in V3 agent. Format: `projects//locations//locations//agents//intents/`. + "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + }, + "parameters": { # In Dialogflow assist for v3, the user can submit a form by sending a SuggestionInput. The form is uniquely determined by the answer_record field, which identifies a v3 QueryResult containing the current page. The form parameters are specified via the parameters field. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "textOverride": { # Represents the natural language text to be processed. # Optional. If the customer edited the suggestion before using it, include the revised text here. + "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. + "text": "A String", # Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters for virtual agent interactions. + }, + }, "textInput": { # Represents the natural language text to be processed. # The natural language text to be processed. "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "text": "A String", # Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters for virtual agent interactions. @@ -1681,62 +1696,1426 @@

Method Details

"contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. }, - "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. - "faqAnswers": [ # Output only. Answers extracted from FAQ documents. - { # Represents answer from "frequently asked questions". - "answer": "A String", # The piece of text from the `source` knowledge base document. - "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "confidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query, range from 0.0 (completely uncertain) to 1.0 (completely certain). - "metadata": { # A map that contains metadata about the answer and the document from which it originates. - "a_key": "A String", - }, - "question": "A String", # The corresponding FAQ question. - "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//locations//agent/knowledgeBases//documents/`. - }, - ], - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. - }, - "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY. - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation. - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. - "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig. - { # Represents a smart reply answer. - "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain). - "reply": "A String", # The content of the reply. - }, - ], - }, - }, - ], - "humanAgentSuggestionResults": [ # The suggestions for most recent human agent. The order is the same as HumanAgentAssistantConfig.SuggestionConfig.feature_configs of HumanAgentAssistantConfig.human_agent_suggestion_config. Note that any failure of Agent Assist features will not lead to the overall failure of an AnalyzeContent API call. Instead, the features will fail silently with the error field set in the corresponding SuggestionResult. - { # One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent. - "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Error status if the request failed. - "code": 42, # The status code, which should be an enum value of google.rpc.Code. - "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. - { - "a_key": "", # Properties of the object. Contains field @type with type URL. - }, - ], - "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. - }, - "suggestArticlesResponse": { # The response message for Participants.SuggestArticles. # SuggestArticlesResponse if request is for ARTICLE_SUGGESTION. - "articleAnswers": [ # Output only. Articles ordered by score in descending order. - { # Represents article answer. + "suggestDialogflowAssistsResponse": { # The response message for Participants.SuggestDialogflowAssists. # SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation. + "dialogflowAssistAnswers": [ # Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction. + { # Represents a Dialogflow assist answer. "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" - "metadata": { # A map that contains metadata about the answer and the document from which it originates. - "a_key": "A String", + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. }, - "snippets": [ # Output only. Article snippets. - "A String", - ], - "title": "A String", # The article title. - "uri": "A String", # The article URI. - }, - ], - "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. - "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`. + }, + "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. + "faqAnswers": [ # Output only. Answers extracted from FAQ documents. + { # Represents answer from "frequently asked questions". + "answer": "A String", # The piece of text from the `source` knowledge base document. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "confidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query, range from 0.0 (completely uncertain) to 1.0 (completely certain). + "metadata": { # A map that contains metadata about the answer and the document from which it originates. + "a_key": "A String", + }, + "question": "A String", # The corresponding FAQ question. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//locations//agent/knowledgeBases//documents/`. + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + }, + "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation. + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig. + { # Represents a smart reply answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain). + "reply": "A String", # The content of the reply. + }, + ], + }, + }, + ], + "humanAgentSuggestionResults": [ # The suggestions for most recent human agent. The order is the same as HumanAgentAssistantConfig.SuggestionConfig.feature_configs of HumanAgentAssistantConfig.human_agent_suggestion_config. Note that any failure of Agent Assist features will not lead to the overall failure of an AnalyzeContent API call. Instead, the features will fail silently with the error field set in the corresponding SuggestionResult. + { # One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Error status if the request failed. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "suggestArticlesResponse": { # The response message for Participants.SuggestArticles. # SuggestArticlesResponse if request is for ARTICLE_SUGGESTION. + "articleAnswers": [ # Output only. Articles ordered by score in descending order. + { # Represents article answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "metadata": { # A map that contains metadata about the answer and the document from which it originates. + "a_key": "A String", + }, + "snippets": [ # Output only. Article snippets. + "A String", + ], + "title": "A String", # The article title. + "uri": "A String", # The article URI. + }, + ], + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestArticlesResponse.context_size field in the request if there aren't that many messages in the conversation. + "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`. + }, + "suggestDialogflowAssistsResponse": { # The response message for Participants.SuggestDialogflowAssists. # SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST. + "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation. + "dialogflowAssistAnswers": [ # Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction. + { # Represents a Dialogflow assist answer. + "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/" + "intentSuggestion": { # Represents an intent suggestion. # An intent suggestion generated from conversation. + "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. + "displayName": "A String", # The display name of the intent. + "intentV2": "A String", # The unique identifier of this intent. Format: `projects//locations//agent/intents/`. + }, + "queryResult": { # Represents the result of conversational query or event processing. # Result from v2 agent. + "action": "A String", # The action name from the matched intent. + "allRequiredParamsPresent": True or False, # This field is set to: - `false` if the matched intent has required parameters and not all of the required parameter values have been collected. - `true` if all required parameter values have been collected, or if the matched intent doesn't contain any required parameters. + "cancelsSlotFilling": True or False, # Indicates whether the conversational query triggers a cancellation for slot filling. For more information, see the [cancel slot filling documentation](https://cloud.google.com/dialogflow/es/docs/intents-actions-parameters#cancel). + "diagnosticInfo": { # Free-form diagnostic information for the associated detect intent request. The fields of this data can change without notice, so you should not write code that depends on its structure. The data may contain: - webhook call latency - webhook errors + "a_key": "", # Properties of the object. + }, + "fulfillmentMessages": [ # The collection of rich messages to present to the user. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "fulfillmentText": "A String", # The text to be pronounced to the user or shown on the screen. Note: This is a legacy field, `fulfillment_messages` should be preferred. + "intent": { # An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification. For more information, see the [intent guide](https://cloud.google.com/dialogflow/docs/intents-overview). # The intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name`, `display_name`, `end_interaction` and `is_fallback`. + "action": "A String", # Optional. The name of the action associated with the intent. Note: The action name must not contain whitespaces. + "defaultResponsePlatforms": [ # Optional. The list of platforms for which the first responses will be copied from the messages in PLATFORM_UNSPECIFIED (i.e. default platform). + "A String", + ], + "displayName": "A String", # Required. The name of this intent. + "endInteraction": True or False, # Optional. Indicates that this intent ends an interaction. Some integrations (e.g., Actions on Google or Dialogflow phone gateway) use this information to close interaction with an end user. Default is false. + "events": [ # Optional. The collection of event names that trigger the intent. If the collection of input contexts is not empty, all of the contexts must be present in the active user session for an event to trigger this intent. Event names are limited to 150 characters. + "A String", + ], + "followupIntentInfo": [ # Output only. Information about all followup intents that have this intent as a direct or indirect parent. We populate this field only in the output. + { # Represents a single followup intent in the chain. + "followupIntentName": "A String", # The unique identifier of the followup intent. Format: `projects//agent/intents/`. + "parentFollowupIntentName": "A String", # The unique identifier of the followup intent's parent. Format: `projects//agent/intents/`. + }, + ], + "inputContextNames": [ # Optional. The list of context names required for this intent to be triggered. Formats: - `projects//agent/sessions/-/contexts/` - `projects//locations//agent/sessions/-/contexts/` + "A String", + ], + "isFallback": True or False, # Optional. Indicates whether this is a fallback intent. + "liveAgentHandoff": True or False, # Optional. Indicates that a live agent should be brought in to handle the interaction with the user. In most cases, when you set this flag to true, you would also want to set end_interaction to true as well. Default is false. + "messages": [ # Optional. The collection of rich messages corresponding to the `Response` field in the Dialogflow console. + { # Corresponds to the `Response` field in the Dialogflow console. + "basicCard": { # The basic card message. Useful for displaying information. # Displays a basic card for Actions on Google. + "buttons": [ # Optional. The collection of card buttons. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "formattedText": "A String", # Required, unless image is present. The body text of the card. + "image": { # The image response message. # Optional. The image for the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "browseCarouselCard": { # Browse Carousel Card for Actions on Google. https://developers.google.com/actions/assistant/responses#browsing_carousel # Browse carousel card for Actions on Google. + "imageDisplayOptions": "A String", # Optional. Settings for displaying the image. Applies to every image in items. + "items": [ # Required. List of items in the Browse Carousel Card. Minimum of two items, maximum of ten. + { # Browsing carousel tile + "description": "A String", # Optional. Description of the carousel item. Maximum of four lines of text. + "footer": "A String", # Optional. Text that appears at the bottom of the Browse Carousel Card. Maximum of one line of text. + "image": { # The image response message. # Optional. Hero image for the carousel item. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "openUriAction": { # Actions on Google action to open a given url. # Required. Action to present to the user. + "url": "A String", # Required. URL + "urlTypeHint": "A String", # Optional. Specifies the type of viewer that is used when opening the URL. Defaults to opening via web browser. + }, + "title": "A String", # Required. Title of the carousel item. Maximum of two lines of text. + }, + ], + }, + "card": { # The card response message. # Displays a card. + "buttons": [ # Optional. The collection of card buttons. + { # Optional. Contains information about a button. + "postback": "A String", # Optional. The text to send back to the Dialogflow API or a URI to open. + "text": "A String", # Optional. The text to show on the button. + }, + ], + "imageUri": "A String", # Optional. The public URI to an image file for the card. + "subtitle": "A String", # Optional. The subtitle of the card. + "title": "A String", # Optional. The title of the card. + }, + "carouselSelect": { # The card for presenting a carousel of options to select from. # Displays a carousel card for Actions on Google. + "items": [ # Required. Carousel items. + { # An item in the carousel. + "description": "A String", # Optional. The body text of the card. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional info about the option item. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. Title of the carousel item. + }, + ], + }, + "image": { # The image response message. # Displays an image. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "linkOutSuggestion": { # The suggestion chip message that allows the user to jump out to the app or website associated with this agent. # Displays a link out suggestion chip for Actions on Google. + "destinationName": "A String", # Required. The name of the app or site this chip is linking to. + "uri": "A String", # Required. The URI of the app or site to open when the user taps the suggestion chip. + }, + "listSelect": { # The card for presenting a list of options to select from. # Displays a list card for Actions on Google. + "items": [ # Required. List items. + { # An item in the list. + "description": "A String", # Optional. The main text describing the item. + "image": { # The image response message. # Optional. The image to display. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "info": { # Additional info about the select item for when it is triggered in a dialog. # Required. Additional information about this option. + "key": "A String", # Required. A unique key that will be sent back to the agent if this response is given. + "synonyms": [ # Optional. A list of synonyms that can also be used to trigger this item in dialog. + "A String", + ], + }, + "title": "A String", # Required. The title of the list item. + }, + ], + "subtitle": "A String", # Optional. Subtitle of the list. + "title": "A String", # Optional. The overall title of the list. + }, + "mediaContent": { # The media content card for Actions on Google. # The media content card for Actions on Google. + "mediaObjects": [ # Required. List of media objects. + { # Response media object for media content card. + "contentUrl": "A String", # Required. Url where the media is stored. + "description": "A String", # Optional. Description of media card. + "icon": { # The image response message. # Optional. Icon to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "largeImage": { # The image response message. # Optional. Image to display above media content. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "name": "A String", # Required. Name of media card. + }, + ], + "mediaType": "A String", # Optional. What type of media is the content (ie "audio"). + }, + "payload": { # A custom platform-specific response. + "a_key": "", # Properties of the object. + }, + "platform": "A String", # Optional. The platform that this message is intended for. + "quickReplies": { # The quick replies response message. # Displays quick replies. + "quickReplies": [ # Optional. The collection of quick replies. + "A String", + ], + "title": "A String", # Optional. The title of the collection of quick replies. + }, + "rbmCarouselRichCard": { # Carousel Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. If you want to show a single card with more control over the layout, please use RbmStandaloneCard instead. # Rich Business Messaging (RBM) carousel rich card response. + "cardContents": [ # Required. The cards in the carousel. A carousel must have at least 2 cards and at most 10. + { # Rich Business Messaging (RBM) Card content + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + ], + "cardWidth": "A String", # Required. The width of the cards in the carousel. + }, + "rbmStandaloneRichCard": { # Standalone Rich Business Messaging (RBM) rich card. Rich cards allow you to respond to users with more vivid content, e.g. with media and suggestions. You can group multiple rich cards into one using RbmCarouselCard but carousel cards will give you less control over the card layout. # Standalone Rich Business Messaging (RBM) rich card response. + "cardContent": { # Rich Business Messaging (RBM) Card content # Required. Card content. + "description": "A String", # Optional. Description of the card (at most 2000 bytes). At least one of the title, description or media must be set. + "media": { # Rich Business Messaging (RBM) Media displayed in Cards The following media-types are currently supported: Image Types * image/jpeg * image/jpg' * image/gif * image/png Video Types * video/h263 * video/m4v * video/mp4 * video/mpeg * video/mpeg4 * video/webm # Optional. However at least one of the title, description or media must be set. Media (image, GIF or a video) to include in the card. + "fileUri": "A String", # Required. Publicly reachable URI of the file. The RBM platform determines the MIME type of the file from the content-type field in the HTTP headers when the platform fetches the file. The content-type field must be present and accurate in the HTTP response from the URL. + "height": "A String", # Required for cards with vertical orientation. The height of the media within a rich card with a vertical layout. For a standalone card with horizontal layout, height is not customizable, and this field is ignored. + "thumbnailUri": "A String", # Optional. Publicly reachable URI of the thumbnail.If you don't provide a thumbnail URI, the RBM platform displays a blank placeholder thumbnail until the user's device downloads the file. Depending on the user's setting, the file may not download automatically and may require the user to tap a download button. + }, + "suggestions": [ # Optional. List of suggestions to include in the card. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "title": "A String", # Optional. Title of the card (at most 200 bytes). At least one of the title, description or media must be set. + }, + "cardOrientation": "A String", # Required. Orientation of the card. + "thumbnailImageAlignment": "A String", # Required if orientation is horizontal. Image preview alignment for standalone cards with horizontal layout. + }, + "rbmText": { # Rich Business Messaging (RBM) text response with suggestions. # Rich Business Messaging (RBM) text response. RBM allows businesses to send enriched and branded versions of SMS. See https://jibe.google.com/business-messaging. + "rbmSuggestion": [ # Optional. One or more suggestions to show to the user. + { # Rich Business Messaging (RBM) suggestion. Suggestions allow user to easily select/click a predefined response or perform an action (like opening a web uri). + "action": { # Rich Business Messaging (RBM) suggested client-side action that the user can choose from the card. # Predefined client side actions that user can choose + "dial": { # Opens the user's default dialer app with the specified phone number but does not dial automatically. # Suggested client side action: Dial a phone number + "phoneNumber": "A String", # Required. The phone number to fill in the default dialer app. This field should be in [E.164](https://en.wikipedia.org/wiki/E.164) format. An example of a correctly formatted phone number: +15556767888. + }, + "openUrl": { # Opens the user's default web browser app to the specified uri If the user has an app installed that is registered as the default handler for the URL, then this app will be opened instead, and its icon will be used in the suggested action UI. # Suggested client side action: Open a URI on device + "uri": "A String", # Required. The uri to open on the user device + }, + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested action. This data will be also forwarded to webhook to allow performing custom business logic. + "shareLocation": { # Opens the device's location chooser so the user can pick a location to send back to the agent. # Suggested client side action: Share user location + }, + "text": "A String", # Text to display alongside the action. + }, + "reply": { # Rich Business Messaging (RBM) suggested reply that the user can click instead of typing in their own response. # Predefined replies for user to select instead of typing + "postbackData": "A String", # Opaque payload that the Dialogflow receives in a user event when the user taps the suggested reply. This data will be also forwarded to webhook to allow performing custom business logic. + "text": "A String", # Suggested reply text. + }, + }, + ], + "text": "A String", # Required. Text sent and displayed to the user. + }, + "simpleResponses": { # The collection of simple response candidates. This message in `QueryResult.fulfillment_messages` and `WebhookResponse.fulfillment_messages` should contain only one `SimpleResponse`. # Returns a voice or text-only response for Actions on Google. + "simpleResponses": [ # Required. The list of simple responses. + { # The simple response message containing speech or text. + "displayText": "A String", # Optional. The text to display. + "ssml": "A String", # One of text_to_speech or ssml must be provided. Structured spoken response to the user in the SSML format. Mutually exclusive with text_to_speech. + "textToSpeech": "A String", # One of text_to_speech or ssml must be provided. The plain text of the speech output. Mutually exclusive with ssml. + }, + ], + }, + "suggestions": { # The collection of suggestions. # Displays suggestion chips for Actions on Google. + "suggestions": [ # Required. The list of suggested replies. + { # The suggestion chip message that the user can tap to quickly post a reply to the conversation. + "title": "A String", # Required. The text shown the in the suggestion chip. + }, + ], + }, + "tableCard": { # Table card for Actions on Google. # Table card for Actions on Google. + "buttons": [ # Optional. List of buttons for the card. + { # The button object that appears at the bottom of a card. + "openUriAction": { # Opens the given URI. # Required. Action to take when a user taps on the button. + "uri": "A String", # Required. The HTTP or HTTPS scheme URI. + }, + "title": "A String", # Required. The title of the button. + }, + ], + "columnProperties": [ # Optional. Display properties for the columns in this table. + { # Column properties for TableCard. + "header": "A String", # Required. Column heading. + "horizontalAlignment": "A String", # Optional. Defines text alignment for all cells in this column. + }, + ], + "image": { # The image response message. # Optional. Image which should be displayed on the card. + "accessibilityText": "A String", # A text description of the image to be used for accessibility, e.g., screen readers. Required if image_uri is set for CarouselSelect. + "imageUri": "A String", # Optional. The public URI to an image file. + }, + "rows": [ # Optional. Rows in this table of data. + { # Row of TableCard. + "cells": [ # Optional. List of cells that make up this row. + { # Cell of TableCardRow. + "text": "A String", # Required. Text in this cell. + }, + ], + "dividerAfter": True or False, # Optional. Whether to add a visual divider after this row. + }, + ], + "subtitle": "A String", # Optional. Subtitle to the title. + "title": "A String", # Required. Title of the card. + }, + "telephonyPlayAudio": { # Plays audio from a file in Telephony Gateway. # Plays audio from a file in Telephony Gateway. + "audioUri": "A String", # Required. URI to a Google Cloud Storage object containing the audio to play, e.g., "gs://bucket/object". The object must contain a single channel (mono) of linear PCM audio (2 bytes / sample) at 8kHz. This object must be readable by the `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` service account where is the number of the Telephony Gateway project (usually the same as the Dialogflow agent project). If the Google Cloud Storage bucket is in the Telephony Gateway project, this permission is added by default when enabling the Dialogflow V2 API. For audio from other sources, consider using the `TelephonySynthesizeSpeech` message with SSML. + }, + "telephonySynthesizeSpeech": { # Synthesizes speech and plays back the synthesized audio to the caller in Telephony Gateway. Telephony Gateway takes the synthesizer settings from `DetectIntentResponse.output_audio_config` which can either be set at request-level or can come from the agent-level synthesizer config. # Synthesizes speech in Telephony Gateway. + "ssml": "A String", # The SSML to be synthesized. For more information, see [SSML](https://developers.google.com/actions/reference/ssml). + "text": "A String", # The raw text to be synthesized. + }, + "telephonyTransferCall": { # Transfers the call in Telephony Gateway. # Transfers the call in Telephony Gateway. + "phoneNumber": "A String", # Required. The phone number to transfer the call to in [E.164 format](https://en.wikipedia.org/wiki/E.164). We currently only allow transferring to US numbers (+1xxxyyyzzzz). + }, + "text": { # The text response message. # Returns a text response. + "text": [ # Optional. The collection of the agent's responses. + "A String", + ], + }, + }, + ], + "mlDisabled": True or False, # Optional. Indicates whether Machine Learning is disabled for the intent. Note: If `ml_disabled` setting is set to true, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. + "mlEnabled": True or False, # Optional. Indicates whether Machine Learning is enabled for the intent. Note: If `ml_enabled` setting is set to false, then this intent is not taken into account during inference in `ML ONLY` match mode. Also, auto-markup in the UI is turned off. DEPRECATED! Please use `ml_disabled` field instead. NOTE: If both `ml_enabled` and `ml_disabled` are either not set or false, then the default value is determined as follows: - Before April 15th, 2018 the default is: ml_enabled = false / ml_disabled = true. - After April 15th, 2018 the default is: ml_enabled = true / ml_disabled = false. + "name": "A String", # Optional. The unique identifier of this intent. Required for Intents.UpdateIntent and Intents.BatchUpdateIntents methods. Supported formats: - `projects//agent/intents/` - `projects//locations//agent/intents/` + "outputContexts": [ # Optional. The collection of contexts that are activated when the intent is matched. Context messages in this collection should not set the parameters field. Setting the `lifespan_count` to 0 will reset the context when the intent is matched. Format: `projects//agent/sessions/-/contexts/`. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": [ # Optional. The collection of parameters associated with the intent. + { # Represents intent parameters. + "defaultValue": "A String", # Optional. The default value to use when the `value` yields an empty result. Default values can be extracted from contexts by using the following syntax: `#context_name.parameter_name`. + "displayName": "A String", # Required. The name of the parameter. + "entityTypeDisplayName": "A String", # Optional. The name of the entity type, prefixed with `@`, that describes values of the parameter. If the parameter is required, this must be provided. + "isList": True or False, # Optional. Indicates whether the parameter represents a list of values. + "mandatory": True or False, # Optional. Indicates whether the parameter is required. That is, whether the intent cannot be completed without collecting the parameter value. + "name": "A String", # The unique identifier of this parameter. + "prompts": [ # Optional. The collection of prompts that the agent can present to the user in order to collect a value for the parameter. + "A String", + ], + "value": "A String", # Optional. The definition of the parameter value. It can be: - a constant string, - a parameter value defined as `$parameter_name`, - an original parameter value defined as `$parameter_name.original`, - a parameter value from some context defined as `#context_name.parameter_name`. + }, + ], + "parentFollowupIntentName": "A String", # Optional. The unique identifier of the parent intent in the chain of followup intents. You can set this field when creating an intent, for example with CreateIntent or BatchUpdateIntents, in order to make this intent a followup intent. It identifies the parent followup intent. Format: `projects//agent/intents/`. + "priority": 42, # Optional. The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. + "resetContexts": True or False, # Optional. Indicates whether to delete all contexts in the current session when this intent is matched. + "rootFollowupIntentName": "A String", # Output only. The unique identifier of the root intent in the chain of followup intents. It identifies the correct followup intents chain for this intent. Format: `projects//agent/intents/`. + "trainingPhrases": [ # Optional. The collection of examples that the agent is trained on. + { # Represents an example that the agent is trained on. + "name": "A String", # Output only. The unique identifier of this training phrase. + "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `entity_type`, `alias`, and `user_defined` fields are all set. + { # Represents a part of a training phrase. + "alias": "A String", # Optional. The parameter name for the value extracted from the annotated part of the example. This field is required for annotated parts of the training phrase. + "entityType": "A String", # Optional. The entity type name prefixed with `@`. This field is required for annotated parts of the training phrase. + "text": "A String", # Required. The text for this part. + "userDefined": True or False, # Optional. Indicates whether the text was manually annotated. This field is set to true when the Dialogflow Console is used to manually annotate the part. When creating an annotated part with the API, you must set this to true. + }, + ], + "timesAddedCount": 42, # Optional. Indicates how many times this example was added to the intent. Each time a developer adds an existing sample by editing an intent or training, this counter is increased. + "type": "A String", # Required. The type of the training phrase. + }, + ], + "webhookState": "A String", # Optional. Indicates whether webhooks are enabled for the intent. + }, + "intentDetectionConfidence": 3.14, # The intent detection confidence. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. If there are `multiple knowledge_answers` messages, this value is set to the greatest `knowledgeAnswers.match_confidence` value in the list. + "knowledgeAnswers": { # Represents the result of querying a Knowledge base. # The result from Knowledge Connector (if any), ordered by decreasing `KnowledgeAnswers.match_confidence`. + "answers": [ # A list of answers from Knowledge Connector. + { # An answer from Knowledge Connector. + "answer": "A String", # The piece of text from the `source` knowledge base document that answers this conversational query. + "faqQuestion": "A String", # The corresponding FAQ question if the answer was extracted from a FAQ Document, empty otherwise. + "matchConfidence": 3.14, # The system's confidence score that this Knowledge answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). Note: The confidence score is likely to vary somewhat (possibly even for identical requests), as the underlying model is under constant improvement. It may be deprecated in the future. We recommend using `match_confidence_level` which should be generally more stable. + "matchConfidenceLevel": "A String", # The system's confidence level that this knowledge answer is a good match for this conversational query. NOTE: The confidence level for a given `` pair may change without notice, as it depends on models that are constantly being improved. However, it will change less frequently than the confidence score below, and should be preferred for referencing the quality of an answer. + "source": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. + }, + ], + }, + "languageCode": "A String", # The language that was triggered during intent detection. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "outputContexts": [ # The collection of output contexts. If applicable, `output_contexts.parameters` contains entries with name `.original` containing the original parameter values before the query. + { # Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview). + "lifespanCount": 42, # Optional. The number of conversational query requests after which the context expires. The default is `0`. If set to `0`, the context expires immediately. Contexts expire automatically after 20 minutes if there are no matching queries. + "name": "A String", # Required. The unique identifier of the context. Supported formats: - `projects//agent/sessions//contexts/`, - `projects//locations//agent/sessions//contexts/`, - `projects//agent/environments//users//sessions//contexts/`, - `projects//locations//agent/environments//users//sessions//contexts/`, The `Context ID` is always converted to lowercase, may only contain characters in a-zA-Z0-9_-% and may be at most 250 bytes long. If `Environment ID` is not specified, we assume default 'draft' environment. If `User ID` is not specified, we assume default '-' user. The following context names are reserved for internal use by Dialogflow. You should not use these contexts or create contexts with these names: * `__system_counters__` * `*_id_dialog_context` * `*_dialog_params_size` + "parameters": { # Optional. The collection of parameters associated with this context. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + }, + ], + "parameters": { # The collection of extracted parameters. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value + "a_key": "", # Properties of the object. + }, + "queryText": "A String", # The original conversational query text: - If natural language text was provided as input, `query_text` contains a copy of the input. - If natural language speech audio was provided as input, `query_text` contains the speech recognition result. If speech recognizer produced multiple alternatives, a particular one is picked. - If automatic spell correction is enabled, `query_text` will contain the corrected user input. + "sentimentAnalysisResult": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # The sentiment analysis result, which depends on the `sentiment_analysis_request_config` specified in the request. + "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`. + "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative). + "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment). + }, + }, + "speechRecognitionConfidence": 3.14, # The Speech recognition confidence between 0.0 and 1.0. A higher number indicates an estimated greater likelihood that the recognized words are correct. The default of 0.0 is a sentinel value indicating that confidence was not set. This field is not guaranteed to be accurate or set. In particular this field isn't set for StreamingDetectIntent since the streaming endpoint has separate confidence estimates per portion of the audio in StreamingRecognitionResult. + "webhookPayload": { # If the query was fulfilled by a webhook call, this field is set to the value of the `payload` field returned in the webhook response. + "a_key": "", # Properties of the object. + }, + "webhookSource": "A String", # If the query was fulfilled by a webhook call, this field is set to the value of the `source` field returned in the webhook response. + }, + }, + ], + "latestMessage": "A String", # The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`. }, "suggestFaqAnswersResponse": { # The request message for Participants.SuggestFaqAnswers. # SuggestFaqAnswersResponse if request is for FAQ_ANSWER. "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestFaqAnswersRequest.context_size field in the request if there aren't that many messages in the conversation. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.suggestions.html index 98dec8eb427..17aa41c04a2 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.suggestions.html @@ -96,11 +96,6 @@

Method Details

The object takes the form of: { # The request message for Conversations.SuggestConversationSummary. - "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. - "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` - "a_key": "A String", - }, - }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000. "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. If empty, the latest message of the conversation will be used. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.html b/docs/dyn/dialogflow_v2beta1.projects.locations.html index 8613ba40e35..94406a0d441 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.html @@ -104,6 +104,11 @@

Instance Methods

Returns the operations Resource.

+

+ suggestions() +

+

Returns the suggestions Resource.

+

close()

Close httplib2 connections.

@@ -163,7 +168,7 @@

Method Details

Returns: An object of the form: - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", @@ -227,7 +232,7 @@

Method Details

{ # The response message for Locations.ListLocations. "locations": [ # A list of locations that matches the specified filter in the request. - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html new file mode 100644 index 00000000000..6c4a737dd62 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html @@ -0,0 +1,306 @@ + + + +

Dialogflow API . projects . locations . suggestions

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ generateStatelessSummary(parent, body=None, x__xgafv=None)

+

Generates and returns a summary for a conversation that does not have a resource created for it.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ generateStatelessSummary(parent, body=None, x__xgafv=None) +
Generates and returns a summary for a conversation that does not have a resource created for it.
+
+Args:
+  parent: string, Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # The request message for Conversations.GenerateStatelessSummary.
+  "conversationProfile": { # Defines the services to connect to incoming Dialogflow conversations. # Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}
+    "automatedAgentConfig": { # Defines the Automated Agent to connect to a conversation. # Configuration for an automated agent to use with this profile.
+      "agent": "A String", # Required. ID of the Dialogflow agent environment to use. This project needs to either be the same project as the conversation or you need to grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow API Service Agent` role in this project. - For ES agents, use format: `projects//locations//agent/environments/`. If environment is not specified, the default `draft` environment is used. Refer to [DetectIntentRequest](/dialogflow/docs/reference/rpc/google.cloud.dialogflow.v2beta1#google.cloud.dialogflow.v2beta1.DetectIntentRequest) for more details. - For CX agents, use format `projects//locations//agents//environments/`. If environment is not specified, the default `draft` environment is used.
+    },
+    "createTime": "A String", # Output only. Create time of the conversation profile.
+    "displayName": "A String", # Required. Human readable name for this profile. Max length 1024 bytes.
+    "humanAgentAssistantConfig": { # Defines the Human Agent Assistant to connect to a conversation. # Configuration for agent assistance to use with this profile.
+      "endUserSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of end user participant. Currently, this feature is not general available, please contact Google to get access.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+                "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent.
+                  "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`.
+                },
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, only one knowledge base is supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION, FAQ, and DIALOGFLOW_ASSIST will use this field.
+              "noSmallTalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+                "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent.
+                  "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`.
+                },
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, only one knowledge base is supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION, FAQ, and DIALOGFLOW_ASSIST will use this field.
+              "noSmallTalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis.
+        "enableEntityExtraction": True or False, # Enable entity extraction in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Currently, this feature is not general available, please contact Google to get access.
+        "enableSentimentAnalysis": True or False, # Enable sentiment analysis in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral: https://cloud.google.com/natural-language/docs/basics#sentiment_analysis For Participants.StreamingAnalyzeContent method, result will be in StreamingAnalyzeContentResponse.message.SentimentAnalysisResult. For Participants.AnalyzeContent method, result will be in AnalyzeContentResponse.message.SentimentAnalysisResult For Conversations.ListMessages method, result will be in ListMessagesResponse.messages.SentimentAnalysisResult If Pub/Sub notification is configured, result will be in ConversationEvent.new_message_payload.SentimentAnalysisResult.
+      },
+      "notificationConfig": { # Defines notification behavior. # Pub/Sub topic on which to publish new agent assistant events.
+        "messageFormat": "A String", # Format of message.
+        "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+      },
+    },
+    "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access.
+      "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com).
+        "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page.
+      },
+      "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent.
+        "buttonId": "A String", # Required. Live Agent chat button ID.
+        "deploymentId": "A String", # Required. Live Agent deployment ID.
+        "endpointDomain": "A String", # Required. Domain of the Live Agent endpoint for this agent. You can find the endpoint URL in the `Live Agent settings` page. For example if URL has the form https://d.la4-c2-phx.salesforceliveagent.com/..., you should fill in d.la4-c2-phx.salesforceliveagent.com.
+        "organizationId": "A String", # Required. The organization ID of the Salesforce account.
+      },
+    },
+    "languageCode": "A String", # Language code for the conversation profile. If not specified, the language is en-US. Language at ConversationProfile should be set for all non en-us languages. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+    "loggingConfig": { # Defines logging behavior for conversation lifecycle events. # Configuration for logging conversation lifecycle events.
+      "enableStackdriverLogging": True or False, # Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.
+    },
+    "name": "A String", # The unique identifier of this conversation profile. Format: `projects//locations//conversationProfiles/`.
+    "newMessageEventNotificationConfig": { # Defines notification behavior. # Configuration for publishing new message events. Event will be sent in format of ConversationEvent
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "notificationConfig": { # Defines notification behavior. # Configuration for publishing conversation lifecycle events.
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
+    "sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
+      "model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then a default model is used. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details.
+      "speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
+    },
+    "timeZone": "A String", # The time zone of this conversational profile from the [time zone database](https://www.iana.org/time-zones), e.g., America/New_York, Europe/Paris. Defaults to America/New_York.
+    "ttsConfig": { # Configuration of how speech should be synthesized. # Configuration for Text-to-Speech synthesization. Used by Phone Gateway to specify synthesization options. If agent defines synthesization options as well, agent settings overrides the option here.
+      "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
+        "A String",
+      ],
+      "pitch": 3.14, # Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
+      "speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
+      "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio.
+        "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices).
+        "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
+      },
+      "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
+    },
+    "updateTime": "A String", # Output only. Update time of the conversation profile.
+  },
+  "latestMessage": "A String", # The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.
+  "maxContextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.
+  "statelessConversation": { # The minimum amount of information required to generate a Summary without having a Conversation resource created. # Required. The conversation to suggest a summary for.
+    "messages": [ # Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.
+      { # Represents a message posted into a conversation.
+        "content": "A String", # Required. The message content.
+        "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
+        "languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+        "messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
+          "containEntities": True or False, # Required. Indicates whether the text message contains entities.
+          "parts": [ # Optional. The collection of annotated message parts ordered by their position in the message. You can recover the annotated message by concatenating [AnnotatedMessagePart.text].
+            { # Represents a part of a message possibly annotated with an entity. The part can be an entity or purely a part of the message between two entities or message start/end.
+              "entityType": "A String", # Optional. The [Dialogflow system entity type](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. If this is empty, Dialogflow could not annotate the phrase part with a system entity.
+              "formattedValue": "", # Optional. The [Dialogflow system entity formatted value ](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. For example for a system entity of type `@sys.unit-currency`, this may contain: { "amount": 5, "currency": "USD" }
+              "text": "A String", # Required. A part of a message possibly annotated with an entity.
+            },
+          ],
+        },
+        "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+        "participant": "A String", # Output only. The participant that sends this message.
+        "participantRole": "A String", # Output only. The role of the participant.
+        "sendTime": "A String", # Optional. The time when the message was sent.
+        "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+          "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`.
+            "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+            "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+          },
+        },
+      },
+    ],
+  },
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # The response message for Conversations.GenerateStatelessSummary.
+  "contextSize": 42, # Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.
+  "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.
+  "summary": { # Generated summary for a conversation. # Generated summary.
+    "text": "A String", # The summary content that is concatenated into one string.
+    "textSections": { # The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.
+      "a_key": "A String",
+    },
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2beta1.projects.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html new file mode 100644 index 00000000000..a3fb24eedd1 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html @@ -0,0 +1,306 @@ + + + +

Dialogflow API . projects . suggestions

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ generateStatelessSummary(parent, body=None, x__xgafv=None)

+

Generates and returns a summary for a conversation that does not have a resource created for it.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ generateStatelessSummary(parent, body=None, x__xgafv=None) +
Generates and returns a summary for a conversation that does not have a resource created for it.
+
+Args:
+  parent: string, Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # The request message for Conversations.GenerateStatelessSummary.
+  "conversationProfile": { # Defines the services to connect to incoming Dialogflow conversations. # Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}
+    "automatedAgentConfig": { # Defines the Automated Agent to connect to a conversation. # Configuration for an automated agent to use with this profile.
+      "agent": "A String", # Required. ID of the Dialogflow agent environment to use. This project needs to either be the same project as the conversation or you need to grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow API Service Agent` role in this project. - For ES agents, use format: `projects//locations//agent/environments/`. If environment is not specified, the default `draft` environment is used. Refer to [DetectIntentRequest](/dialogflow/docs/reference/rpc/google.cloud.dialogflow.v2beta1#google.cloud.dialogflow.v2beta1.DetectIntentRequest) for more details. - For CX agents, use format `projects//locations//agents//environments/`. If environment is not specified, the default `draft` environment is used.
+    },
+    "createTime": "A String", # Output only. Create time of the conversation profile.
+    "displayName": "A String", # Required. Human readable name for this profile. Max length 1024 bytes.
+    "humanAgentAssistantConfig": { # Defines the Human Agent Assistant to connect to a conversation. # Configuration for agent assistance to use with this profile.
+      "endUserSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of end user participant. Currently, this feature is not general available, please contact Google to get access.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+                "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent.
+                  "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`.
+                },
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, only one knowledge base is supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION, FAQ, and DIALOGFLOW_ASSIST will use this field.
+              "noSmallTalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant.
+        "featureConfigs": [ # Configuration of different suggestion features. One feature can have only one config.
+          { # Config for suggestion features.
+            "conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION. # Configs of custom conversation model.
+              "baselineModelVersion": "A String", # Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0
+              "model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
+            },
+            "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+              "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+            },
+            "enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
+            "queryConfig": { # Config for suggestion query. # Configs of query.
+              "confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.
+              "contextFilterSettings": { # Settings that determine how to filter recent conversation context when generating suggestions. # Determines how recent conversation context is filtered when generating suggestions. If unspecified, no messages will be dropped.
+                "dropHandoffMessages": True or False, # If set to true, the last message from virtual agent (hand off message) and the message before it (trigger message of hand off) are dropped.
+                "dropIvrMessages": True or False, # If set to true, all messages from ivr stage are dropped.
+                "dropVirtualAgentMessages": True or False, # If set to true, all messages from virtual agent are dropped.
+              },
+              "dialogflowQuerySource": { # Dialogflow source setting. Supported feature: DIALOGFLOW_ASSIST. # Query from Dialogflow agent. It is used by DIALOGFLOW_ASSIST.
+                "agent": "A String", # Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.
+                "humanAgentSideConfig": { # The configuration used for human agent side Dialogflow assist suggestion. # The Dialogflow assist configuration for human agent.
+                  "agent": "A String", # Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`.
+                },
+              },
+              "documentQuerySource": { # Document source settings. Supported features: SMART_REPLY, SMART_COMPOSE. # Query from knowledge base document. It is used by: SMART_REPLY, SMART_COMPOSE.
+                "documents": [ # Required. Knowledge documents to query from. Format: `projects//locations//knowledgeBases//documents/`. Currently, only one document is supported.
+                  "A String",
+                ],
+              },
+              "knowledgeBaseQuerySource": { # Knowledge base source settings. Supported features: ARTICLE_SUGGESTION, FAQ. # Query from knowledgebase. It is used by: ARTICLE_SUGGESTION, FAQ.
+                "knowledgeBases": [ # Required. Knowledge bases to query. Format: `projects//locations//knowledgeBases/`. Currently, only one knowledge base is supported.
+                  "A String",
+                ],
+              },
+              "maxResults": 42, # Maximum number of results to return. Currently, if unset, defaults to 10. And the max number is 20.
+            },
+            "suggestionFeature": { # The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple `Feature` objects can be specified in the `features` list. # The suggestion feature.
+              "type": "A String", # Type of Human Agent Assistant API feature to request.
+            },
+            "suggestionTriggerSettings": { # Settings of suggestion trigger. # Settings of suggestion trigger. Currently, only ARTICLE_SUGGESTION, FAQ, and DIALOGFLOW_ASSIST will use this field.
+              "noSmallTalk": True or False, # Do not trigger if last utterance is small talk.
+              "onlyEndUser": True or False, # Only trigger suggestion if participant role of last utterance is END_USER.
+            },
+          },
+        ],
+        "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
+      },
+      "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis.
+        "enableEntityExtraction": True or False, # Enable entity extraction in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Currently, this feature is not general available, please contact Google to get access.
+        "enableSentimentAnalysis": True or False, # Enable sentiment analysis in conversation messages on [agent assist stage](https://cloud.google.com/dialogflow/priv/docs/contact-center/basics#stages). If unspecified, defaults to false. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral: https://cloud.google.com/natural-language/docs/basics#sentiment_analysis For Participants.StreamingAnalyzeContent method, result will be in StreamingAnalyzeContentResponse.message.SentimentAnalysisResult. For Participants.AnalyzeContent method, result will be in AnalyzeContentResponse.message.SentimentAnalysisResult For Conversations.ListMessages method, result will be in ListMessagesResponse.messages.SentimentAnalysisResult If Pub/Sub notification is configured, result will be in ConversationEvent.new_message_payload.SentimentAnalysisResult.
+      },
+      "notificationConfig": { # Defines notification behavior. # Pub/Sub topic on which to publish new agent assistant events.
+        "messageFormat": "A String", # Format of message.
+        "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+      },
+    },
+    "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access.
+      "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com).
+        "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page.
+      },
+      "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent.
+        "buttonId": "A String", # Required. Live Agent chat button ID.
+        "deploymentId": "A String", # Required. Live Agent deployment ID.
+        "endpointDomain": "A String", # Required. Domain of the Live Agent endpoint for this agent. You can find the endpoint URL in the `Live Agent settings` page. For example if URL has the form https://d.la4-c2-phx.salesforceliveagent.com/..., you should fill in d.la4-c2-phx.salesforceliveagent.com.
+        "organizationId": "A String", # Required. The organization ID of the Salesforce account.
+      },
+    },
+    "languageCode": "A String", # Language code for the conversation profile. If not specified, the language is en-US. Language at ConversationProfile should be set for all non en-us languages. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+    "loggingConfig": { # Defines logging behavior for conversation lifecycle events. # Configuration for logging conversation lifecycle events.
+      "enableStackdriverLogging": True or False, # Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.
+    },
+    "name": "A String", # The unique identifier of this conversation profile. Format: `projects//locations//conversationProfiles/`.
+    "newMessageEventNotificationConfig": { # Defines notification behavior. # Configuration for publishing new message events. Event will be sent in format of ConversationEvent
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "notificationConfig": { # Defines notification behavior. # Configuration for publishing conversation lifecycle events.
+      "messageFormat": "A String", # Format of message.
+      "topic": "A String", # Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant `service-@gcp-sa-dialogflow.iam.gserviceaccount.com` the `Dialogflow Service Agent` role in the topic project. For chat integration to receive notification, make sure API caller has been granted the `Dialogflow Service Agent` role for the topic. Format: `projects//locations//topics/`.
+    },
+    "securitySettings": "A String", # Name of the CX SecuritySettings reference for the agent. Format: `projects//locations//securitySettings/`.
+    "sttConfig": { # Configures speech transcription for ConversationProfile. # Settings for speech transcription.
+      "model": "A String", # Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then a default model is used. Refer to [Cloud Speech API documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) for more details.
+      "speechModelVariant": "A String", # The speech model used in speech to text. `SPEECH_MODEL_VARIANT_UNSPECIFIED`, `USE_BEST_AVAILABLE` will be treated as `USE_ENHANCED`. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
+    },
+    "timeZone": "A String", # The time zone of this conversational profile from the [time zone database](https://www.iana.org/time-zones), e.g., America/New_York, Europe/Paris. Defaults to America/New_York.
+    "ttsConfig": { # Configuration of how speech should be synthesized. # Configuration for Text-to-Speech synthesization. Used by Phone Gateway to specify synthesization options. If agent defines synthesization options as well, agent settings overrides the option here.
+      "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
+        "A String",
+      ],
+      "pitch": 3.14, # Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
+      "speakingRate": 3.14, # Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
+      "voice": { # Description of which voice to use for speech synthesis. # Optional. The desired voice of the synthesized audio.
+        "name": "A String", # Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to [Supported voices and languages](https://cloud.google.com/text-to-speech/docs/voices).
+        "ssmlGender": "A String", # Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
+      },
+      "volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
+    },
+    "updateTime": "A String", # Output only. Update time of the conversation profile.
+  },
+  "latestMessage": "A String", # The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.
+  "maxContextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.
+  "statelessConversation": { # The minimum amount of information required to generate a Summary without having a Conversation resource created. # Required. The conversation to suggest a summary for.
+    "messages": [ # Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.
+      { # Represents a message posted into a conversation.
+        "content": "A String", # Required. The message content.
+        "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
+        "languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
+        "messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
+          "containEntities": True or False, # Required. Indicates whether the text message contains entities.
+          "parts": [ # Optional. The collection of annotated message parts ordered by their position in the message. You can recover the annotated message by concatenating [AnnotatedMessagePart.text].
+            { # Represents a part of a message possibly annotated with an entity. The part can be an entity or purely a part of the message between two entities or message start/end.
+              "entityType": "A String", # Optional. The [Dialogflow system entity type](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. If this is empty, Dialogflow could not annotate the phrase part with a system entity.
+              "formattedValue": "", # Optional. The [Dialogflow system entity formatted value ](https://cloud.google.com/dialogflow/docs/reference/system-entities) of this message part. For example for a system entity of type `@sys.unit-currency`, this may contain: { "amount": 5, "currency": "USD" }
+              "text": "A String", # Required. A part of a message possibly annotated with an entity.
+            },
+          ],
+        },
+        "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+        "participant": "A String", # Output only. The participant that sends this message.
+        "participantRole": "A String", # Output only. The role of the participant.
+        "sendTime": "A String", # Optional. The time when the message was sent.
+        "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+          "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. See: https://cloud.google.com/natural-language/docs/basics#interpreting_sentiment_analysis_values for how to interpret the result. # The sentiment analysis result for `query_text`.
+            "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+            "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+          },
+        },
+      },
+    ],
+  },
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # The response message for Conversations.GenerateStatelessSummary.
+  "contextSize": 42, # Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.
+  "latestMessage": "A String", # The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.
+  "summary": { # Generated summary for a conversation. # Generated summary.
+    "text": "A String", # The summary content that is concatenated into one string.
+    "textSections": { # The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.
+      "a_key": "A String",
+    },
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.html b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.html index 3e75aae0c82..74de8c56471 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.html @@ -168,12 +168,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -181,12 +187,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -325,12 +337,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -338,12 +356,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -398,12 +422,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -411,12 +441,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -488,12 +524,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -501,12 +543,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -569,12 +617,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -582,12 +636,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html index 91f9e317d2e..10cb51dd6a5 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.environments.sessions.html @@ -123,7 +123,7 @@

Method Details

"volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. }, }, - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -858,7 +858,7 @@

Method Details

"digits": "A String", # The dtmf digits. "finishDigit": "A String", # The finish digit (if any). }, - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -893,7 +893,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # Intent match result, could be an intent or an event. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1022,7 +1022,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # The matched intent/event to fulfill. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1060,7 +1060,7 @@

Method Details

}, "matchIntentRequest": { # Request of MatchIntent. # Must be same as the corresponding MatchIntent request, otherwise the behavior is undefined. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1811,7 +1811,7 @@

Method Details

"digits": "A String", # The dtmf digits. "finishDigit": "A String", # The finish digit (if any). }, - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1846,7 +1846,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # Intent match result, could be an intent or an event. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1972,7 +1972,7 @@

Method Details

{ # Request of MatchIntent. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -2685,7 +2685,7 @@

Method Details

{ # Represents one match result of MatchIntent. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.html b/docs/dyn/dialogflow_v3.projects.locations.agents.html index 7470e4d429e..c2e2b875d5c 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.html @@ -165,7 +165,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -188,8 +188,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -218,7 +218,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -241,8 +241,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -340,7 +340,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -363,8 +363,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -446,7 +446,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -469,8 +469,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -518,7 +518,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -541,8 +541,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -572,7 +572,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -595,8 +595,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.intents.html b/docs/dyn/dialogflow_v3.projects.locations.agents.intents.html index caa50261ac1..1fbd2491e32 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.intents.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.intents.html @@ -110,7 +110,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 +{ # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -150,7 +150,7 @@

Method Details

Returns: An object of the form: - { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 + { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -215,7 +215,7 @@

Method Details

Returns: An object of the form: - { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 + { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -271,7 +271,7 @@

Method Details

{ # The response message for Intents.ListIntents. "intents": [ # The list of intents. There will be a maximum number of items returned based on the page_size field in the request. - { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 + { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -329,7 +329,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 +{ # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -370,7 +370,7 @@

Method Details

Returns: An object of the form: - { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 + { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html index c8dca2f1a1f..0bbd21095ea 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.sessions.html @@ -123,7 +123,7 @@

Method Details

"volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. }, }, - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -858,7 +858,7 @@

Method Details

"digits": "A String", # The dtmf digits. "finishDigit": "A String", # The finish digit (if any). }, - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -893,7 +893,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # Intent match result, could be an intent or an event. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1022,7 +1022,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # The matched intent/event to fulfill. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1060,7 +1060,7 @@

Method Details

}, "matchIntentRequest": { # Request of MatchIntent. # Must be same as the corresponding MatchIntent request, otherwise the behavior is undefined. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1811,7 +1811,7 @@

Method Details

"digits": "A String", # The dtmf digits. "finishDigit": "A String", # The finish digit (if any). }, - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the conversational query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. This field is deprecated, please use QueryResult.match instead. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1846,7 +1846,7 @@

Method Details

"match": { # Represents one match result of MatchIntent. # Intent match result, could be an intent or an event. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -1972,7 +1972,7 @@

Method Details

{ # Request of MatchIntent. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -2685,7 +2685,7 @@

Method Details

{ # Represents one match result of MatchIntent. "confidence": 3.14, # The confidence of this match. Values range from 0.0 (completely uncertain) to 1.0 (completely certain). This value is for informational purpose only and is only used to help match the best intent within the classification threshold. This value may change for the same end-user expression at any time due to a model retraining or change in implementation. "event": "A String", # The event that matched the query. Filled for `EVENT`, `NO_MATCH` and `NO_INPUT` match types. - "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. + "intent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that matched the query. Some, not all fields are filled in this message, including but not limited to: `name` and `display_name`. Only filled for `INTENT` match type. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html index 33f9de4bb49..866578ce25a 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.html @@ -214,7 +214,7 @@

Method Details

"coverageScore": 3.14, # The percent of intents in the agent that are covered. "intents": [ # The list of Intents present in the agent { # The agent's intent. - "covered": True or False, # Whether or not the intent is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the intent is covered by at least one of the agent's test cases. "intent": "A String", # The intent full resource name }, ], @@ -356,7 +356,7 @@

Method Details

}, "transitions": [ # The list of transition routes and coverage in the transition route group. { # A transition coverage in a transition route group. - "covered": True or False, # Whether or not the transition route is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the transition route is covered by at least one of the agent's test cases. "transitionRoute": { # A transition route specifies a intent that can be matched and/or a data condition that can be evaluated during a session. When a specified transition is matched, the following actions are taken in order: * If there is a `trigger_fulfillment` associated with the transition, it will be called. * If there is a `target_page` associated with the transition, the session will transition into the specified page. * If there is a `target_flow` associated with the transition, the session will transition into the specified flow. # Intent route or condition route. "condition": "A String", # The condition to evaluate against form parameters or session parameters. See the [conditions reference](https://cloud.google.com/dialogflow/cx/docs/reference/condition). At least one of `intent` or `condition` must be specified. When both `intent` and `condition` are specified, the transition can only happen when both are fulfilled. "intent": "A String", # The unique identifier of an Intent. Format: `projects//locations//agents//intents/`. Indicates that the transition can only happen when the given intent is matched. At least one of `intent` or `condition` must be specified. When both `intent` and `condition` are specified, the transition can only happen when both are fulfilled. @@ -490,7 +490,7 @@

Method Details

"coverageScore": 3.14, # The percent of transitions in the agent that are covered. "transitions": [ # The list of Transitions present in the agent. { # A transition in a page. - "covered": True or False, # Whether or not the transition is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the transition is covered by at least one of the agent's test cases. "eventHandler": { # An event handler specifies an event that can be handled during a session. When the specified event happens, the following actions are taken in order: * If there is a `trigger_fulfillment` associated with the event, it will be called. * If there is a `target_page` associated with the event, the session will transition into the specified page. * If there is a `target_flow` associated with the event, the session will transition into the specified flow. # Event handler. "event": "A String", # Required. The name of the event to handle. "name": "A String", # Output only. The unique identifier of this event handler. @@ -2564,7 +2564,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -3257,7 +3257,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -3307,7 +3307,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -4000,7 +4000,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -4061,7 +4061,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -4754,7 +4754,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -4804,7 +4804,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -5497,7 +5497,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -5609,7 +5609,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -6302,7 +6302,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -6352,7 +6352,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -7045,7 +7045,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -7166,7 +7166,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -7859,7 +7859,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -7909,7 +7909,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -8602,7 +8602,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -8681,7 +8681,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -9374,7 +9374,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -9424,7 +9424,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -10117,7 +10117,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -10179,7 +10179,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -10872,7 +10872,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -10922,7 +10922,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -11615,7 +11615,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html index 02cc531ff82..95aed885d4d 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.testCases.results.html @@ -114,7 +114,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -807,7 +807,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. @@ -875,7 +875,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1568,7 +1568,7 @@

Method Details

], }, ], - "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15 # The Intent that triggered the response. Only name and displayName will be set. + "triggeredIntent": { # An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. # The Intent that triggered the response. Only name and displayName will be set. "description": "A String", # Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters. "displayName": "A String", # Required. The human-readable name of the intent, unique within the agent. "isFallback": True or False, # Indicates whether this is a fallback intent. Currently only default fallback intent is allowed in the agent, which is added upon agent creation. Adding training phrases to fallback intent is useful in the case of requests that are mistakenly matched, since training phrases assigned to fallback intents act as negative examples that triggers no-match event. diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html b/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html index 2246b075a96..00ae048f975 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html @@ -117,12 +117,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -130,12 +136,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -157,12 +169,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -170,12 +188,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -223,12 +247,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -236,12 +266,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -275,12 +311,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -288,12 +330,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -333,12 +381,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -346,12 +400,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -374,12 +434,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -387,12 +453,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, diff --git a/docs/dyn/dialogflow_v3.projects.locations.html b/docs/dyn/dialogflow_v3.projects.locations.html index e64d87ca08b..172cf049f72 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.html +++ b/docs/dyn/dialogflow_v3.projects.locations.html @@ -121,7 +121,7 @@

Method Details

Returns: An object of the form: - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", @@ -153,7 +153,7 @@

Method Details

{ # The response message for Locations.ListLocations. "locations": [ # A list of locations that matches the specified filter in the request. - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.html index 60ead153ce7..3b475cd213d 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.html @@ -168,12 +168,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -181,12 +187,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -325,12 +337,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -338,12 +356,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -398,12 +422,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -411,12 +441,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -488,12 +524,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -501,12 +543,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -569,12 +617,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -582,12 +636,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html index 7fddc773e0e..1dd1a734071 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.environments.sessions.html @@ -123,7 +123,7 @@

Method Details

"volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. }, }, - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -876,7 +876,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -911,7 +911,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1040,7 +1040,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1060,7 +1060,7 @@

Method Details

}, "matchIntentRequest": { # Request of MatchIntent. # Must be same as the corresponding MatchIntent request, otherwise the behavior is undefined. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1829,7 +1829,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1864,7 +1864,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1972,7 +1972,7 @@

Method Details

{ # Request of MatchIntent. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -2703,7 +2703,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html index 7df1c890fd6..04ccddecdd1 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.html @@ -165,7 +165,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -188,8 +188,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -218,7 +218,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -241,8 +241,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -340,7 +340,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -363,8 +363,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -446,7 +446,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -469,8 +469,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -518,7 +518,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -541,8 +541,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", @@ -572,7 +572,7 @@

Method Details

{ # Agents are best described as Natural Language Understanding (NLU) modules that transform user requests into actionable data. You can include agents in your app, product, or service to determine user intent and respond to the user in a natural way. After you create an agent, you can add Intents, Entity Types, Flows, Fulfillments, Webhooks, and so on to manage the conversation flows.. "advancedSettings": { # Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter. # Hierarchical advanced settings for this agent. The settings exposed at the lower level overrides the settings exposed at the higher level. "audioExportGcsDestination": { # Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow. # If present, incoming audio is exported by Dialogflow to the configured Google Cloud Storage destination. Exposed at the following levels: - Agent level - Flow level - "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. + "uri": "A String", # Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation. }, "loggingSettings": { # Define behaviors on logging. # Settings for logging. Settings for Dialogflow History, Contact Center messages, StackDriver logs, and speech logging. Exposed at the following levels: - Agent level. "enableInteractionLogging": True or False, # If true, DF Interaction logging is currently enabled. @@ -595,8 +595,8 @@

Method Details

"supportedLanguageCodes": [ # The list of all languages supported by the agent (except for the `default_language_code`). "A String", ], - "textToSpeechSettings": { # Settings related to speech generating. # Settings on instructing the speech synthesizer on how to generate the output audio content. - "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig. + "textToSpeechSettings": { # Settings related to speech synthesizing. # Settings on instructing the speech synthesizer on how to generate the output audio content. + "synthesizeSpeechConfigs": { # Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing. "a_key": { # Configuration of how speech should be synthesized. "effectsProfileId": [ # Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given. "A String", diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.intents.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.intents.html index 96d263554cd..1e266ebff90 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.intents.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.intents.html @@ -128,7 +128,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -168,7 +168,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -233,7 +233,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -289,7 +289,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -347,7 +347,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -388,7 +388,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html index ae2c917fc7c..ea866199e58 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.sessions.html @@ -123,7 +123,7 @@

Method Details

"volumeGainDb": 3.14, # Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that. }, }, - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -876,7 +876,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -911,7 +911,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1040,7 +1040,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1060,7 +1060,7 @@

Method Details

}, "matchIntentRequest": { # Request of MatchIntent. # Must be same as the corresponding MatchIntent request, otherwise the behavior is undefined. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1829,7 +1829,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1864,7 +1864,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -1972,7 +1972,7 @@

Method Details

{ # Request of MatchIntent. "persistParameterChanges": True or False, # Persist session parameter changes from `query_params`. - "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Required. The input specification. + "queryInput": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Required. The input specification. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -2703,7 +2703,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html index c7b948cd6db..c88bcfd149f 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.html @@ -214,7 +214,7 @@

Method Details

"coverageScore": 3.14, # The percent of intents in the agent that are covered. "intents": [ # The list of Intents present in the agent { # The agent's intent. - "covered": True or False, # Whether or not the intent is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the intent is covered by at least one of the agent's test cases. "intent": "A String", # The intent full resource name }, ], @@ -356,7 +356,7 @@

Method Details

}, "transitions": [ # The list of transition routes and coverage in the transition route group. { # A transition coverage in a transition route group. - "covered": True or False, # Whether or not the transition route is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the transition route is covered by at least one of the agent's test cases. "transitionRoute": { # A transition route specifies a intent that can be matched and/or a data condition that can be evaluated during a session. When a specified transition is matched, the following actions are taken in order: * If there is a `trigger_fulfillment` associated with the transition, it will be called. * If there is a `target_page` associated with the transition, the session will transition into the specified page. * If there is a `target_flow` associated with the transition, the session will transition into the specified flow. # Intent route or condition route. "condition": "A String", # The condition to evaluate against form parameters or session parameters. See the [conditions reference](https://cloud.google.com/dialogflow/cx/docs/reference/condition). At least one of `intent` or `condition` must be specified. When both `intent` and `condition` are specified, the transition can only happen when both are fulfilled. "intent": "A String", # The unique identifier of an Intent. Format: `projects//locations//agents//intents/`. Indicates that the transition can only happen when the given intent is matched. At least one of `intent` or `condition` must be specified. When both `intent` and `condition` are specified, the transition can only happen when both are fulfilled. @@ -490,7 +490,7 @@

Method Details

"coverageScore": 3.14, # The percent of transitions in the agent that are covered. "transitions": [ # The list of Transitions present in the agent. { # A transition in a page. - "covered": True or False, # Whether or not the transition is covered by at least one of the agent's test cases. + "covered": True or False, # Whether the transition is covered by at least one of the agent's test cases. "eventHandler": { # An event handler specifies an event that can be handled during a session. When the specified event happens, the following actions are taken in order: * If there is a `trigger_fulfillment` associated with the event, it will be called. * If there is a `target_page` associated with the event, the session will transition into the specified page. * If there is a `target_flow` associated with the event, the session will transition into the specified flow. # Event handler. "event": "A String", # Required. The name of the event to handle. "name": "A String", # Output only. The unique identifier of this event handler. @@ -2564,7 +2564,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -3275,7 +3275,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -3307,7 +3307,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -4018,7 +4018,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -4061,7 +4061,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -4772,7 +4772,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -4804,7 +4804,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -5515,7 +5515,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -5609,7 +5609,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -6320,7 +6320,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -6352,7 +6352,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -7063,7 +7063,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -7166,7 +7166,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -7877,7 +7877,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -7909,7 +7909,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -8620,7 +8620,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -8681,7 +8681,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -9392,7 +9392,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -9424,7 +9424,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -10135,7 +10135,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -10179,7 +10179,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -10890,7 +10890,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -10922,7 +10922,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -11633,7 +11633,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html index 36ab05ba6c5..7bab94aa3d3 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.testCases.results.html @@ -114,7 +114,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -825,7 +825,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. @@ -875,7 +875,7 @@

Method Details

"injectedParameters": { # Parameters that need to be injected into the conversation during intent detection. "a_key": "", # Properties of the object. }, - "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. # Supports text input, event input, dtmf input in the test case. + "input": { # Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value. # Supports text input, event input, dtmf input in the test case. "audio": { # Represents the natural speech audio to be processed. # The natural language speech audio to be processed. "audio": "A String", # The natural language speech audio to be processed. A single request can contain up to 2 minutes of speech audio data. The transcribed text cannot contain more than 256 bytes. For non-streaming audio detect intent, both `config` and `audio` must be provided. For streaming audio detect intent, `config` must be provided in the first request and `audio` must be provided in all following requests. "config": { # Instructs the speech recognizer on how to process the audio content. # Required. Instructs the speech recognizer how to process the speech audio. @@ -1586,7 +1586,7 @@

Method Details

], "priority": 42, # The priority of this intent. Higher numbers represent higher priorities. - If the supplied value is unspecified or 0, the service translates the value to 500,000, which corresponds to the `Normal` priority in the console. - If the supplied value is negative, the intent is ignored in runtime detect intent requests. "trainingPhrases": [ # The collection of training phrases the agent is trained on to identify the intent. - { # Represents an example that the agent is trained on to identify the intent. Next ID: 15 + { # Represents an example that the agent is trained on to identify the intent. "id": "A String", # Output only. The unique identifier of the training phrase. "parts": [ # Required. The ordered list of training phrase parts. The parts are concatenated in order to form the training phrase. Note: The API does not automatically annotate training phrases like the Dialogflow Console does. Note: Do not forget to include whitespace at part boundaries, so the training phrase is well formatted when the parts are concatenated. If the training phrase does not need to be annotated with parameters, you just need a single part with only the Part.text field set. If you want to annotate the training phrase, you must create multiple parts, where the fields of each part are populated in one of two ways: - `Part.text` is set to a part of the phrase that has no parameters. - `Part.text` is set to a part of the phrase that you want to annotate, and the `parameter_id` field is set. { # Represents a part of a training phrase. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html index 66ad1e90de6..3dab8916da5 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html @@ -117,12 +117,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -130,12 +136,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -157,12 +169,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -170,12 +188,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -223,12 +247,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -236,12 +266,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -275,12 +311,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -288,12 +330,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -333,12 +381,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -346,12 +400,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, @@ -374,12 +434,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. @@ -387,12 +453,18 @@

Method Details

"allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, ``` openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") ``` "A String", ], + "httpMethod": "A String", # Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST. + "parameterMapping": { # Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response + "a_key": "A String", + }, "password": "A String", # The password for HTTP Basic authentication. + "requestBody": "A String", # Optional. Defines a custom JSON object as request body to send to flexible webhook. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", }, "uri": "A String", # Required. The webhook URI for receiving POST requests. It must use https protocol. "username": "A String", # The user name for HTTP Basic authentication. + "webhookType": "A String", # Optional. Type of the webhook. }, "service": "A String", # Required. The name of [Service Directory](https://cloud.google.com/service-directory) service. Format: `projects//locations//namespaces//services/`. `Location ID` of the service directory must be the same as the location of the agent. }, diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.html b/docs/dyn/dialogflow_v3beta1.projects.locations.html index 7b533b7e8f2..2d9446be517 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.html @@ -121,7 +121,7 @@

Method Details

Returns: An object of the form: - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", @@ -153,7 +153,7 @@

Method Details

{ # The response message for Locations.ListLocations. "locations": [ # A list of locations that matches the specified filter in the request. - { # A resource that represents Google Cloud Platform location. + { # A resource that represents a Google Cloud location. "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} "a_key": "A String", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2.json b/googleapiclient/discovery_cache/documents/dialogflow.v2.json index 53666c98e6c..29c6964ebac 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2.json @@ -8039,6 +8039,39 @@ ] } } + }, + "suggestions": { + "methods": { + "generateStatelessSummary": { + "description": "Generates and returns a summary for a conversation that does not have a resource created for it.", + "flatPath": "v2/projects/{projectsId}/locations/{locationsId}/suggestions:generateStatelessSummary", + "httpMethod": "POST", + "id": "dialogflow.projects.locations.suggestions.generateStatelessSummary", + "parameterOrder": [ + "parent" + ], + "parameters": { + "parent": { + "description": "Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`.", + "location": "path", + "pattern": "^projects/[^/]+/locations/[^/]+$", + "required": true, + "type": "string" + } + }, + "path": "v2/{+parent}/suggestions:generateStatelessSummary", + "request": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryRequest" + }, + "response": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryResponse" + }, + "scopes": [ + "https://www.googleapis.com/auth/cloud-platform", + "https://www.googleapis.com/auth/dialogflow" + ] + } + } } } }, @@ -8139,11 +8172,44 @@ ] } } + }, + "suggestions": { + "methods": { + "generateStatelessSummary": { + "description": "Generates and returns a summary for a conversation that does not have a resource created for it.", + "flatPath": "v2/projects/{projectsId}/suggestions:generateStatelessSummary", + "httpMethod": "POST", + "id": "dialogflow.projects.suggestions.generateStatelessSummary", + "parameterOrder": [ + "parent" + ], + "parameters": { + "parent": { + "description": "Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`.", + "location": "path", + "pattern": "^projects/[^/]+$", + "required": true, + "type": "string" + } + }, + "path": "v2/{+parent}/suggestions:generateStatelessSummary", + "request": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryRequest" + }, + "response": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryResponse" + }, + "scopes": [ + "https://www.googleapis.com/auth/cloud-platform", + "https://www.googleapis.com/auth/dialogflow" + ] + } + } } } } }, - "revision": "20230407", + "revision": "20230516", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AudioInput": { @@ -8543,7 +8609,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -8916,7 +8982,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3Intent": { - "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15", + "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent.", "id": "GoogleCloudDialogflowCxV3Intent", "properties": { "description": { @@ -9153,7 +9219,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3QueryInput", "properties": { "audio": { @@ -9604,14 +9670,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -9702,6 +9770,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -9771,10 +9849,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -9789,6 +9902,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -10408,7 +10535,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -10864,7 +10991,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase": { - "description": "Represents an example that the agent is trained on to identify the intent. Next ID: 15", + "description": "Represents an example that the agent is trained on to identify the intent.", "id": "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase", "properties": { "id": { @@ -11018,7 +11145,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3beta1QueryInput", "properties": { "audio": { @@ -11469,14 +11596,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -11567,6 +11696,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -11636,10 +11775,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -11654,6 +11828,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -12173,7 +12361,7 @@ "type": "string" }, "clicked": { - "description": "Indicates whether the answer/item was clicked by the human agent or not. Default to false.", + "description": "Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search and knowledge assist, the answer record is considered to be clicked if the answer was copied or any URI was clicked.", "type": "boolean" }, "correctnessLevel": { @@ -13740,6 +13928,82 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2GenerateStatelessSummaryRequest": { + "description": "The request message for Conversations.GenerateStatelessSummary.", + "id": "GoogleCloudDialogflowV2GenerateStatelessSummaryRequest", + "properties": { + "conversationProfile": { + "$ref": "GoogleCloudDialogflowV2ConversationProfile", + "description": "Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}" + }, + "latestMessage": { + "description": "The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.", + "type": "string" + }, + "maxContextSize": { + "description": "Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.", + "format": "int32", + "type": "integer" + }, + "statelessConversation": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryRequestMinimalConversation", + "description": "Required. The conversation to suggest a summary for." + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2GenerateStatelessSummaryRequestMinimalConversation": { + "description": "The minimum amount of information required to generate a Summary without having a Conversation resource created.", + "id": "GoogleCloudDialogflowV2GenerateStatelessSummaryRequestMinimalConversation", + "properties": { + "messages": { + "description": "Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.", + "items": { + "$ref": "GoogleCloudDialogflowV2Message" + }, + "type": "array" + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2GenerateStatelessSummaryResponse": { + "description": "The response message for Conversations.GenerateStatelessSummary.", + "id": "GoogleCloudDialogflowV2GenerateStatelessSummaryResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "latestMessage": { + "description": "The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.", + "type": "string" + }, + "summary": { + "$ref": "GoogleCloudDialogflowV2GenerateStatelessSummaryResponseSummary", + "description": "Generated summary." + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2GenerateStatelessSummaryResponseSummary": { + "description": "Generated summary for a conversation.", + "id": "GoogleCloudDialogflowV2GenerateStatelessSummaryResponseSummary", + "properties": { + "text": { + "description": "The summary content that is concatenated into one string.", + "type": "string" + }, + "textSections": { + "additionalProperties": { + "type": "string" + }, + "description": "The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.", + "type": "object" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2HumanAgentAssistantConfig": { "description": "Defines the Human Agent Assist to connect to a conversation.", "id": "GoogleCloudDialogflowV2HumanAgentAssistantConfig", @@ -13767,6 +14031,10 @@ "description": "Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION.", "id": "GoogleCloudDialogflowV2HumanAgentAssistantConfigConversationModelConfig", "properties": { + "baselineModelVersion": { + "description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0", + "type": "string" + }, "model": { "description": "Conversation model resource name. Format: `projects//conversationModels/`.", "type": "string" @@ -13855,7 +14123,7 @@ "id": "GoogleCloudDialogflowV2HumanAgentAssistantConfigSuggestionQueryConfig", "properties": { "confidenceThreshold": { - "description": "Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.", + "description": "Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.", "format": "float", "type": "number" }, @@ -14197,6 +14465,10 @@ "description": "Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.", "type": "boolean" }, + "enableAutomaticPunctuation": { + "description": "Enable automatic punctuation option at the speech backend.", + "type": "boolean" + }, "enableWordInfo": { "description": "If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information.", "type": "boolean" @@ -16305,10 +16577,6 @@ "description": "The request message for Conversations.SuggestConversationSummary.", "id": "GoogleCloudDialogflowV2SuggestConversationSummaryRequest", "properties": { - "assistQueryParams": { - "$ref": "GoogleCloudDialogflowV2AssistQueryParameters", - "description": "Parameters for a human assist query." - }, "contextSize": { "description": "Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.", "format": "int32", @@ -16924,6 +17192,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -16931,6 +17200,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -17001,6 +17271,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer": { + "description": "Represents a Dialogflow assist answer.", + "id": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer", + "properties": { + "answerRecord": { + "description": "The name of answer record, in the format of \"projects//locations//answerRecords/\"", + "type": "string" + }, + "intentSuggestion": { + "$ref": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "description": "An intent suggestion generated from conversation." + }, + "queryResult": { + "$ref": "GoogleCloudDialogflowV2beta1QueryResult", + "description": "Result from v2 agent." + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1EntityType": { "description": "Each intent parameter has a type, called the entity type, which dictates exactly how data from an end-user expression is extracted. Dialogflow provides predefined system entities that can match many common types of data. For example, there are system entities for matching dates, times, colors, email addresses, and so on. You can also create your own custom entities for matching custom data. For example, you could define a vegetable entity that can match the types of vegetables available for purchase with a grocery store agent. For more information, see the [Entity guide](https://cloud.google.com/dialogflow/docs/entities-overview).", "id": "GoogleCloudDialogflowV2beta1EntityType", @@ -18351,6 +18640,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1IntentSuggestion": { + "description": "Represents an intent suggestion.", + "id": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "properties": { + "description": { + "description": "Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters.", + "type": "string" + }, + "displayName": { + "description": "The display name of the intent.", + "type": "string" + }, + "intentV2": { + "description": "The unique identifier of this intent. Format: `projects//locations//agent/intents/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1IntentTrainingPhrase": { "description": "Represents an example that the agent is trained on.", "id": "GoogleCloudDialogflowV2beta1IntentTrainingPhrase", @@ -18791,6 +19099,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -18798,6 +19107,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -18848,6 +19158,29 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse": { + "description": "The response message for Participants.SuggestDialogflowAssists.", + "id": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "dialogflowAssistAnswers": { + "description": "Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction.", + "items": { + "$ref": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer" + }, + "type": "array" + }, + "latestMessage": { + "description": "The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse": { "description": "The request message for Participants.SuggestFaqAnswers.", "id": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", @@ -18906,6 +19239,10 @@ "$ref": "GoogleCloudDialogflowV2beta1SuggestArticlesResponse", "description": "SuggestArticlesResponse if request is for ARTICLE_SUGGESTION." }, + "suggestDialogflowAssistsResponse": { + "$ref": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "description": "SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST." + }, "suggestFaqAnswersResponse": { "$ref": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", "description": "SuggestFaqAnswersResponse if request is for FAQ_ANSWER." @@ -19137,6 +19474,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "triggeredAbandonmentEvent": { "description": "Whether agent has triggered the event corresponding to user abandoning the conversation.", "type": "boolean" @@ -19185,7 +19532,7 @@ "type": "object" }, "GoogleCloudLocationLocation": { - "description": "A resource that represents Google Cloud Platform location.", + "description": "A resource that represents a Google Cloud location.", "id": "GoogleCloudLocationLocation", "properties": { "displayName": { diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json index af8cd2cbcd4..fade3c2769b 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json @@ -2609,6 +2609,7 @@ "answerRecords": { "methods": { "get": { + "deprecated": true, "description": "Deprecated. Retrieves a specific answer record.", "flatPath": "v2beta1/projects/{projectsId}/answerRecords/{answerRecordsId}", "httpMethod": "GET", @@ -3301,6 +3302,7 @@ "suggestions": { "methods": { "compile": { + "deprecated": true, "description": "Deprecated. use SuggestArticles and SuggestFaqAnswers instead. Gets suggestions for a participant based on specific historical messages. Note that ListSuggestions will only list the auto-generated suggestions, while CompileSuggestion will try to compile suggestion based on the provided conversation context in the real time.", "flatPath": "v2beta1/projects/{projectsId}/conversations/{conversationsId}/participants/{participantsId}/suggestions:compile", "httpMethod": "POST", @@ -3330,6 +3332,7 @@ ] }, "list": { + "deprecated": true, "description": "Deprecated: Use inline suggestion, event based suggestion or Suggestion* API instead. See HumanAgentAssistantConfig.name for more details. Removal Date: 2020-09-01. Retrieves suggestions for live agents. This method should be used by human agent client software to fetch auto generated suggestions in real-time, while the conversation with an end user is in progress. The functionality is implemented in terms of the [list pagination](https://cloud.google.com/apis/design/design_patterns#list_pagination) design pattern. The client app should use the `next_page_token` field to fetch the next batch of suggestions. `suggestions` are sorted by `create_time` in descending order. To fetch latest suggestion, just set `page_size` to 1. To fetch new suggestions without duplication, send request with filter `create_time_epoch_microseconds > [first item's create_time of previous request]` and empty page_token.", "flatPath": "v2beta1/projects/{projectsId}/conversations/{conversationsId}/participants/{participantsId}/suggestions", "httpMethod": "GET", @@ -6095,6 +6098,7 @@ "answerRecords": { "methods": { "get": { + "deprecated": true, "description": "Deprecated. Retrieves a specific answer record.", "flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/answerRecords/{answerRecordsId}", "httpMethod": "GET", @@ -7403,6 +7407,39 @@ ] } } + }, + "suggestions": { + "methods": { + "generateStatelessSummary": { + "description": "Generates and returns a summary for a conversation that does not have a resource created for it.", + "flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/suggestions:generateStatelessSummary", + "httpMethod": "POST", + "id": "dialogflow.projects.locations.suggestions.generateStatelessSummary", + "parameterOrder": [ + "parent" + ], + "parameters": { + "parent": { + "description": "Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`.", + "location": "path", + "pattern": "^projects/[^/]+/locations/[^/]+$", + "required": true, + "type": "string" + } + }, + "path": "v2beta1/{+parent}/suggestions:generateStatelessSummary", + "request": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequest" + }, + "response": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponse" + }, + "scopes": [ + "https://www.googleapis.com/auth/cloud-platform", + "https://www.googleapis.com/auth/dialogflow" + ] + } + } } } }, @@ -7503,11 +7540,44 @@ ] } } + }, + "suggestions": { + "methods": { + "generateStatelessSummary": { + "description": "Generates and returns a summary for a conversation that does not have a resource created for it.", + "flatPath": "v2beta1/projects/{projectsId}/suggestions:generateStatelessSummary", + "httpMethod": "POST", + "id": "dialogflow.projects.suggestions.generateStatelessSummary", + "parameterOrder": [ + "parent" + ], + "parameters": { + "parent": { + "description": "Required. The parent resource to charge for the Summary's generation. Format: `projects//locations/`.", + "location": "path", + "pattern": "^projects/[^/]+$", + "required": true, + "type": "string" + } + }, + "path": "v2beta1/{+parent}/suggestions:generateStatelessSummary", + "request": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequest" + }, + "response": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponse" + }, + "scopes": [ + "https://www.googleapis.com/auth/cloud-platform", + "https://www.googleapis.com/auth/dialogflow" + ] + } + } } } } }, - "revision": "20230407", + "revision": "20230516", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AudioInput": { @@ -7907,7 +7977,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -8280,7 +8350,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3Intent": { - "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15", + "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent.", "id": "GoogleCloudDialogflowCxV3Intent", "properties": { "description": { @@ -8517,7 +8587,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3QueryInput", "properties": { "audio": { @@ -8968,14 +9038,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -9066,6 +9138,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -9135,10 +9217,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -9153,6 +9270,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -9772,7 +9903,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -10228,7 +10359,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase": { - "description": "Represents an example that the agent is trained on to identify the intent. Next ID: 15", + "description": "Represents an example that the agent is trained on to identify the intent.", "id": "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase", "properties": { "id": { @@ -10382,7 +10513,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3beta1QueryInput", "properties": { "audio": { @@ -10833,14 +10964,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -10931,6 +11064,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -11000,10 +11143,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -11018,6 +11196,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -13546,6 +13738,11 @@ "description": "Output only. The article suggestion answer.", "readOnly": true }, + "dialogflowAssistAnswer": { + "$ref": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer", + "description": "Output only. The Dialogflow assist answer.", + "readOnly": true + }, "faqAnswer": { "$ref": "GoogleCloudDialogflowV2beta1FaqAnswer", "description": "Output only. The FAQ answer.", @@ -13599,6 +13796,10 @@ "description": "A unique identifier for this request. Restricted to 36 ASCII characters. A random UUID is recommended. This request is only idempotent if a `request_id` is provided.", "type": "string" }, + "suggestionInput": { + "$ref": "GoogleCloudDialogflowV2beta1SuggestionInput", + "description": "An input representing the selection of a suggestion." + }, "textInput": { "$ref": "GoogleCloudDialogflowV2beta1TextInput", "description": "The natural language text to be processed." @@ -13680,7 +13881,7 @@ "type": "string" }, "clicked": { - "description": "Indicates whether the answer/item was clicked by the human agent or not. Default to false.", + "description": "Indicates whether the answer/item was clicked by the human agent or not. Default to false. For knowledge search, the answer record is considered to be clicked if the answer was copied or any URI was clicked.", "type": "boolean" }, "correctnessLevel": { @@ -14129,6 +14330,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -14136,6 +14338,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -14170,6 +14373,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -14177,6 +14381,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -14513,6 +14718,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer": { + "description": "Represents a Dialogflow assist answer.", + "id": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer", + "properties": { + "answerRecord": { + "description": "The name of answer record, in the format of \"projects//locations//answerRecords/\"", + "type": "string" + }, + "intentSuggestion": { + "$ref": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "description": "An intent suggestion generated from conversation." + }, + "queryResult": { + "$ref": "GoogleCloudDialogflowV2beta1QueryResult", + "description": "Result from v2 agent." + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1Document": { "description": "A knowledge document to be used by a KnowledgeBase. For more information, see the [knowledge base guide](https://cloud.google.com/dialogflow/docs/how/knowledge-bases). Note: The `projects.agent.knowledgeBases.documents` resource is deprecated; only use `projects.knowledgeBases.documents`.", "id": "GoogleCloudDialogflowV2beta1Document", @@ -15025,6 +15249,82 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequest": { + "description": "The request message for Conversations.GenerateStatelessSummary.", + "id": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequest", + "properties": { + "conversationProfile": { + "$ref": "GoogleCloudDialogflowV2beta1ConversationProfile", + "description": "Required. A ConversationProfile containing information required for Summary generation. Required fields: {language_code, security_settings} Optional fields: {agent_assistant_config}" + }, + "latestMessage": { + "description": "The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.", + "type": "string" + }, + "maxContextSize": { + "description": "Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.", + "format": "int32", + "type": "integer" + }, + "statelessConversation": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequestMinimalConversation", + "description": "Required. The conversation to suggest a summary for." + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequestMinimalConversation": { + "description": "The minimum amount of information required to generate a Summary without having a Conversation resource created.", + "id": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequestMinimalConversation", + "properties": { + "messages": { + "description": "Required. The messages that the Summary will be generated from. It is expected that this message content is already redacted and does not contain any PII. Required fields: {content, language_code, participant, participant_role} Optional fields: {send_time} If send_time is not provided, then the messages must be provided in chronological order.", + "items": { + "$ref": "GoogleCloudDialogflowV2beta1Message" + }, + "type": "array" + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponse": { + "description": "The response message for Conversations.GenerateStatelessSummary.", + "id": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including last_conversation_message used to compile the suggestion. It may be smaller than the GenerateStatelessSummaryRequest.context_size field in the request if there weren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "latestMessage": { + "description": "The name of the latest conversation message used as context for compiling suggestion. The format is specific to the user and the names of the messages provided.", + "type": "string" + }, + "summary": { + "$ref": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponseSummary", + "description": "Generated summary." + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponseSummary": { + "description": "Generated summary for a conversation.", + "id": "GoogleCloudDialogflowV2beta1GenerateStatelessSummaryResponseSummary", + "properties": { + "text": { + "description": "The summary content that is concatenated into one string.", + "type": "string" + }, + "textSections": { + "additionalProperties": { + "type": "string" + }, + "description": "The summary content that is divided into sections. The key is the section's name and the value is the section's content. There is no specific format for the key or value.", + "type": "object" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfig": { "description": "Defines the Human Agent Assistant to connect to a conversation.", "id": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfig", @@ -15052,6 +15352,10 @@ "description": "Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY, CONVERSATION_SUMMARIZATION.", "id": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigConversationModelConfig", "properties": { + "baselineModelVersion": { + "description": "Version of current baseline model. It will be ignored if model is set. Valid versions are: Article Suggestion baseline model: - 0.9 - 1.0 (default) Summarization baseline model: - 1.0", + "type": "string" + }, "model": { "description": "Conversation model resource name. Format: `projects//conversationModels/`.", "type": "string" @@ -15140,7 +15444,7 @@ "id": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigSuggestionQueryConfig", "properties": { "confidenceThreshold": { - "description": "Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.", + "description": "Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it is default to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE, KNOWLEDGE_SEARCH, KNOWLEDGE_ASSIST, ENTITY_EXTRACTION.", "format": "float", "type": "number" }, @@ -15194,6 +15498,21 @@ "agent": { "description": "Required. The name of a dialogflow virtual agent used for end user side intent detection and suggestion. Format: `projects//locations//agent`. When multiple agents are allowed in the same Dialogflow project.", "type": "string" + }, + "humanAgentSideConfig": { + "$ref": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigSuggestionQueryConfigDialogflowQuerySourceHumanAgentSideConfig", + "description": "The Dialogflow assist configuration for human agent." + } + }, + "type": "object" + }, + "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigSuggestionQueryConfigDialogflowQuerySourceHumanAgentSideConfig": { + "description": "The configuration used for human agent side Dialogflow assist suggestion.", + "id": "GoogleCloudDialogflowV2beta1HumanAgentAssistantConfigSuggestionQueryConfigDialogflowQuerySourceHumanAgentSideConfig", + "properties": { + "agent": { + "description": "Optional. The name of a dialogflow virtual agent used for intent detection and suggestion triggered by human agent. Format: `projects//locations//agent`.", + "type": "string" } }, "type": "object" @@ -15438,6 +15757,10 @@ "description": "Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent.", "type": "boolean" }, + "enableAutomaticPunctuation": { + "description": "Enable automatic punctuation option at the speech backend.", + "type": "boolean" + }, "enableWordInfo": { "description": "If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information.", "type": "boolean" @@ -15675,6 +15998,21 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1IntentInput": { + "description": "Represents the intent to trigger programmatically rather than as a result of natural language processing. The intent input is only used for V3 agent.", + "id": "GoogleCloudDialogflowV2beta1IntentInput", + "properties": { + "intent": { + "description": "Required. The unique identifier of the intent in V3 agent. Format: `projects//locations//locations//agents//intents/`.", + "type": "string" + }, + "languageCode": { + "description": "Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1IntentMessage": { "description": "Corresponds to the `Response` field in the Dialogflow console.", "id": "GoogleCloudDialogflowV2beta1IntentMessage", @@ -16652,6 +16990,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1IntentSuggestion": { + "description": "Represents an intent suggestion.", + "id": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "properties": { + "description": { + "description": "Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters.", + "type": "string" + }, + "displayName": { + "description": "The display name of the intent.", + "type": "string" + }, + "intentV2": { + "description": "The unique identifier of this intent. Format: `projects//locations//agent/intents/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1IntentTrainingPhrase": { "description": "Represents an example that the agent is trained on.", "id": "GoogleCloudDialogflowV2beta1IntentTrainingPhrase", @@ -17757,6 +18114,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -17764,6 +18122,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -17926,10 +18285,6 @@ "description": "The request message for Conversations.SuggestConversationSummary.", "id": "GoogleCloudDialogflowV2beta1SuggestConversationSummaryRequest", "properties": { - "assistQueryParams": { - "$ref": "GoogleCloudDialogflowV2beta1AssistQueryParameters", - "description": "Parameters for a human assist query." - }, "contextSize": { "description": "Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.", "format": "int32", @@ -17984,6 +18339,29 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse": { + "description": "The response message for Participants.SuggestDialogflowAssists.", + "id": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "dialogflowAssistAnswers": { + "description": "Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction.", + "items": { + "$ref": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer" + }, + "type": "array" + }, + "latestMessage": { + "description": "The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1SuggestFaqAnswersRequest": { "description": "The request message for Participants.SuggestFaqAnswers.", "id": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersRequest", @@ -18183,6 +18561,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -18190,6 +18569,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -18197,6 +18577,33 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1SuggestionInput": { + "description": "Represents the selection of a suggestion.", + "id": "GoogleCloudDialogflowV2beta1SuggestionInput", + "properties": { + "answerRecord": { + "description": "Required. The ID of a suggestion selected by the human agent. The suggestion(s) were generated in a previous call to request Dialogflow assist. The format is: `projects//locations//answerRecords/` where is an alphanumeric string.", + "type": "string" + }, + "intentInput": { + "$ref": "GoogleCloudDialogflowV2beta1IntentInput", + "description": "The intent to be triggered on V3 agent." + }, + "parameters": { + "additionalProperties": { + "description": "Properties of the object.", + "type": "any" + }, + "description": "In Dialogflow assist for v3, the user can submit a form by sending a SuggestionInput. The form is uniquely determined by the answer_record field, which identifies a v3 QueryResult containing the current page. The form parameters are specified via the parameters field. Depending on your protocol or client library language, this is a map, associative array, symbol table, dictionary, or JSON object composed of a collection of (MapKey, MapValue) pairs: - MapKey type: string - MapKey value: parameter name - MapValue type: - If parameter's entity type is a composite entity: map - Else: depending on parameter value type, could be one of string, number, boolean, null, list or map - MapValue value: - If parameter's entity type is a composite entity: map from composite entity property names to property values - Else: parameter value", + "type": "object" + }, + "textOverride": { + "$ref": "GoogleCloudDialogflowV2beta1TextInput", + "description": "Optional. If the customer edited the suggestion before using it, include the revised text here." + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1SuggestionResult": { "description": "One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent.", "id": "GoogleCloudDialogflowV2beta1SuggestionResult", @@ -18209,6 +18616,10 @@ "$ref": "GoogleCloudDialogflowV2beta1SuggestArticlesResponse", "description": "SuggestArticlesResponse if request is for ARTICLE_SUGGESTION." }, + "suggestDialogflowAssistsResponse": { + "$ref": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "description": "SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST." + }, "suggestFaqAnswersResponse": { "$ref": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", "description": "SuggestFaqAnswersResponse if request is for FAQ_ANSWER." @@ -18711,6 +19122,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "triggeredAbandonmentEvent": { "description": "Whether agent has triggered the event corresponding to user abandoning the conversation.", "type": "boolean" @@ -18759,7 +19180,7 @@ "type": "object" }, "GoogleCloudLocationLocation": { - "description": "A resource that represents Google Cloud Platform location.", + "description": "A resource that represents a Google Cloud location.", "id": "GoogleCloudLocationLocation", "properties": { "displayName": { diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3.json b/googleapiclient/discovery_cache/documents/dialogflow.v3.json index 00cc8d6eb35..512f968dc3a 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3.json @@ -3820,7 +3820,7 @@ } } }, - "revision": "20230407", + "revision": "20230516", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -4952,7 +4952,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -5324,7 +5324,7 @@ "id": "GoogleCloudDialogflowCxV3GcsDestination", "properties": { "uri": { - "description": "Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation.", + "description": "Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation.", "type": "string" } }, @@ -5536,7 +5536,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3Intent": { - "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15", + "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent.", "id": "GoogleCloudDialogflowCxV3Intent", "properties": { "description": { @@ -5608,7 +5608,7 @@ "id": "GoogleCloudDialogflowCxV3IntentCoverageIntent", "properties": { "covered": { - "description": "Whether or not the intent is covered by at least one of the agent's test cases.", + "description": "Whether the intent is covered by at least one of the agent's test cases.", "type": "boolean" }, "intent": { @@ -6325,7 +6325,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3QueryInput", "properties": { "audio": { @@ -7289,14 +7289,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -7315,14 +7317,14 @@ "type": "object" }, "GoogleCloudDialogflowCxV3TextToSpeechSettings": { - "description": "Settings related to speech generating.", + "description": "Settings related to speech synthesizing.", "id": "GoogleCloudDialogflowCxV3TextToSpeechSettings", "properties": { "synthesizeSpeechConfigs": { "additionalProperties": { "$ref": "GoogleCloudDialogflowCxV3SynthesizeSpeechConfig" }, - "description": "Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig.", + "description": "Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing.", "type": "object" } }, @@ -7358,7 +7360,7 @@ "id": "GoogleCloudDialogflowCxV3TransitionCoverageTransition", "properties": { "covered": { - "description": "Whether or not the transition is covered by at least one of the agent's test cases.", + "description": "Whether the transition is covered by at least one of the agent's test cases.", "type": "boolean" }, "eventHandler": { @@ -7501,7 +7503,7 @@ "id": "GoogleCloudDialogflowCxV3TransitionRouteGroupCoverageCoverageTransition", "properties": { "covered": { - "description": "Whether or not the transition route is covered by at least one of the agent's test cases.", + "description": "Whether the transition route is covered by at least one of the agent's test cases.", "type": "boolean" }, "transitionRoute": { @@ -7552,6 +7554,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -7844,10 +7856,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -7862,6 +7909,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -8481,7 +8542,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -8937,7 +8998,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase": { - "description": "Represents an example that the agent is trained on to identify the intent. Next ID: 15", + "description": "Represents an example that the agent is trained on to identify the intent.", "id": "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase", "properties": { "id": { @@ -9091,7 +9152,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3beta1QueryInput", "properties": { "audio": { @@ -9542,14 +9603,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -9640,6 +9703,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -9709,10 +9782,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -9727,6 +9835,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -12195,6 +12317,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -12202,6 +12325,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -12272,6 +12396,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer": { + "description": "Represents a Dialogflow assist answer.", + "id": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer", + "properties": { + "answerRecord": { + "description": "The name of answer record, in the format of \"projects//locations//answerRecords/\"", + "type": "string" + }, + "intentSuggestion": { + "$ref": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "description": "An intent suggestion generated from conversation." + }, + "queryResult": { + "$ref": "GoogleCloudDialogflowV2beta1QueryResult", + "description": "Result from v2 agent." + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1EntityType": { "description": "Each intent parameter has a type, called the entity type, which dictates exactly how data from an end-user expression is extracted. Dialogflow provides predefined system entities that can match many common types of data. For example, there are system entities for matching dates, times, colors, email addresses, and so on. You can also create your own custom entities for matching custom data. For example, you could define a vegetable entity that can match the types of vegetables available for purchase with a grocery store agent. For more information, see the [Entity guide](https://cloud.google.com/dialogflow/docs/entities-overview).", "id": "GoogleCloudDialogflowV2beta1EntityType", @@ -13622,6 +13765,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1IntentSuggestion": { + "description": "Represents an intent suggestion.", + "id": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "properties": { + "description": { + "description": "Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters.", + "type": "string" + }, + "displayName": { + "description": "The display name of the intent.", + "type": "string" + }, + "intentV2": { + "description": "The unique identifier of this intent. Format: `projects//locations//agent/intents/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1IntentTrainingPhrase": { "description": "Represents an example that the agent is trained on.", "id": "GoogleCloudDialogflowV2beta1IntentTrainingPhrase", @@ -14062,6 +14224,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -14069,6 +14232,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -14119,6 +14283,29 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse": { + "description": "The response message for Participants.SuggestDialogflowAssists.", + "id": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "dialogflowAssistAnswers": { + "description": "Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction.", + "items": { + "$ref": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer" + }, + "type": "array" + }, + "latestMessage": { + "description": "The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse": { "description": "The request message for Participants.SuggestFaqAnswers.", "id": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", @@ -14177,6 +14364,10 @@ "$ref": "GoogleCloudDialogflowV2beta1SuggestArticlesResponse", "description": "SuggestArticlesResponse if request is for ARTICLE_SUGGESTION." }, + "suggestDialogflowAssistsResponse": { + "$ref": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "description": "SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST." + }, "suggestFaqAnswersResponse": { "$ref": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", "description": "SuggestFaqAnswersResponse if request is for FAQ_ANSWER." @@ -14408,6 +14599,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "triggeredAbandonmentEvent": { "description": "Whether agent has triggered the event corresponding to user abandoning the conversation.", "type": "boolean" @@ -14456,7 +14657,7 @@ "type": "object" }, "GoogleCloudLocationLocation": { - "description": "A resource that represents Google Cloud Platform location.", + "description": "A resource that represents a Google Cloud location.", "id": "GoogleCloudLocationLocation", "properties": { "displayName": { diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json index ad814bbc474..8ee9cce4886 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json @@ -3820,7 +3820,7 @@ } } }, - "revision": "20230407", + "revision": "20230516", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AudioInput": { @@ -4220,7 +4220,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -4593,7 +4593,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3Intent": { - "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent. Next ID: 15", + "description": "An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent.", "id": "GoogleCloudDialogflowCxV3Intent", "properties": { "description": { @@ -4830,7 +4830,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3QueryInput", "properties": { "audio": { @@ -5281,14 +5281,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -5379,6 +5381,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -5448,10 +5460,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -5466,6 +5513,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -6817,7 +6878,7 @@ "type": "string" }, "agentUri": { - "description": "The URI to a file containing the exported agent. This field is populated only if `agent_uri` is specified in ExportAgentRequest.", + "description": "The URI to a file containing the exported agent. This field is populated if `agent_uri` is specified in ExportAgentRequest.", "type": "string" } }, @@ -7189,7 +7250,7 @@ "id": "GoogleCloudDialogflowCxV3beta1GcsDestination", "properties": { "uri": { - "description": "Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: gs://bucket/object-name-or-prefix Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation.", + "description": "Required. The Google Cloud Storage URI for the exported objects. A URI is of the form: `gs://bucket/object-name-or-prefix` Whether a full object name, or just a prefix, its usage depends on the Dialogflow operation.", "type": "string" } }, @@ -7473,7 +7534,7 @@ "id": "GoogleCloudDialogflowCxV3beta1IntentCoverageIntent", "properties": { "covered": { - "description": "Whether or not the intent is covered by at least one of the agent's test cases.", + "description": "Whether the intent is covered by at least one of the agent's test cases.", "type": "boolean" }, "intent": { @@ -7518,7 +7579,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase": { - "description": "Represents an example that the agent is trained on to identify the intent. Next ID: 15", + "description": "Represents an example that the agent is trained on to identify the intent.", "id": "GoogleCloudDialogflowCxV3beta1IntentTrainingPhrase", "properties": { "id": { @@ -8190,7 +8251,7 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1QueryInput": { - "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. ", + "description": "Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.", "id": "GoogleCloudDialogflowCxV3beta1QueryInput", "properties": { "audio": { @@ -9154,14 +9215,16 @@ "INTENT", "PAGE", "PARAMETERS", - "UTTERANCE" + "UTTERANCE", + "FLOW" ], "enumDescriptions": [ "Should never be used.", "The intent.", "The page.", "The parameters.", - "The message utterance." + "The message utterance.", + "The flow." ], "type": "string" } @@ -9180,14 +9243,14 @@ "type": "object" }, "GoogleCloudDialogflowCxV3beta1TextToSpeechSettings": { - "description": "Settings related to speech generating.", + "description": "Settings related to speech synthesizing.", "id": "GoogleCloudDialogflowCxV3beta1TextToSpeechSettings", "properties": { "synthesizeSpeechConfigs": { "additionalProperties": { "$ref": "GoogleCloudDialogflowCxV3beta1SynthesizeSpeechConfig" }, - "description": "Configuration of how speech should be synthesized, mapping from language (https://dialogflow.com/docs/reference/language) to SynthesizeSpeechConfig.", + "description": "Configuration of how speech should be synthesized, mapping from language (https://cloud.google.com/dialogflow/cx/docs/reference/language) to SynthesizeSpeechConfig. These settings affect: - The synthesize configuration used in [phone gateway](https://cloud.google.com/dialogflow/cx/docs/concept/integration/phone-gateway). - You no longer need to specify OutputAudioConfig.synthesize_speech_config when invoking API calls. Your agent will use the pre-configured options for speech synthesizing.", "type": "object" } }, @@ -9223,7 +9286,7 @@ "id": "GoogleCloudDialogflowCxV3beta1TransitionCoverageTransition", "properties": { "covered": { - "description": "Whether or not the transition is covered by at least one of the agent's test cases.", + "description": "Whether the transition is covered by at least one of the agent's test cases.", "type": "boolean" }, "eventHandler": { @@ -9366,7 +9429,7 @@ "id": "GoogleCloudDialogflowCxV3beta1TransitionRouteGroupCoverageCoverageTransition", "properties": { "covered": { - "description": "Whether or not the transition route is covered by at least one of the agent's test cases.", + "description": "Whether the transition route is covered by at least one of the agent's test cases.", "type": "boolean" }, "transitionRoute": { @@ -9417,6 +9480,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "userEscalated": { "description": "Whether user was specifically asking for a live agent.", "type": "boolean" @@ -9709,10 +9782,45 @@ }, "type": "array" }, + "httpMethod": { + "description": "Optional. HTTP method for the flexible webhook calls. Standard webhook always uses POST.", + "enum": [ + "HTTP_METHOD_UNSPECIFIED", + "POST", + "GET", + "HEAD", + "PUT", + "DELETE", + "PATCH", + "OPTIONS" + ], + "enumDescriptions": [ + "HTTP method not specified.", + "HTTP POST Method.", + "HTTP GET Method.", + "HTTP HEAD Method.", + "HTTP PUT Method.", + "HTTP DELETE Method.", + "HTTP PATCH Method.", + "HTTP OPTIONS Method." + ], + "type": "string" + }, + "parameterMapping": { + "additionalProperties": { + "type": "string" + }, + "description": "Optional. Maps the values extracted from specific fields of the flexible webhook response into session parameters. - Key: session parameter name - Value: field path in the webhook response", + "type": "object" + }, "password": { "description": "The password for HTTP Basic authentication.", "type": "string" }, + "requestBody": { + "description": "Optional. Defines a custom JSON object as request body to send to flexible webhook.", + "type": "string" + }, "requestHeaders": { "additionalProperties": { "type": "string" @@ -9727,6 +9835,20 @@ "username": { "description": "The user name for HTTP Basic authentication.", "type": "string" + }, + "webhookType": { + "description": "Optional. Type of the webhook.", + "enum": [ + "WEBHOOK_TYPE_UNSPECIFIED", + "STANDARD", + "FLEXIBLE" + ], + "enumDescriptions": [ + "Default value. This value is unused.", + "Represents a standard webhook.", + "Represents a flexible webhook." + ], + "type": "string" } }, "type": "object" @@ -12195,6 +12317,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -12202,6 +12325,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -12272,6 +12396,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer": { + "description": "Represents a Dialogflow assist answer.", + "id": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer", + "properties": { + "answerRecord": { + "description": "The name of answer record, in the format of \"projects//locations//answerRecords/\"", + "type": "string" + }, + "intentSuggestion": { + "$ref": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "description": "An intent suggestion generated from conversation." + }, + "queryResult": { + "$ref": "GoogleCloudDialogflowV2beta1QueryResult", + "description": "Result from v2 agent." + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1EntityType": { "description": "Each intent parameter has a type, called the entity type, which dictates exactly how data from an end-user expression is extracted. Dialogflow provides predefined system entities that can match many common types of data. For example, there are system entities for matching dates, times, colors, email addresses, and so on. You can also create your own custom entities for matching custom data. For example, you could define a vegetable entity that can match the types of vegetables available for purchase with a grocery store agent. For more information, see the [Entity guide](https://cloud.google.com/dialogflow/docs/entities-overview).", "id": "GoogleCloudDialogflowV2beta1EntityType", @@ -13622,6 +13765,25 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1IntentSuggestion": { + "description": "Represents an intent suggestion.", + "id": "GoogleCloudDialogflowV2beta1IntentSuggestion", + "properties": { + "description": { + "description": "Human readable description for better understanding an intent like its scope, content, result etc. Maximum character limit: 140 characters.", + "type": "string" + }, + "displayName": { + "description": "The display name of the intent.", + "type": "string" + }, + "intentV2": { + "description": "The unique identifier of this intent. Format: `projects//locations//agent/intents/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1IntentTrainingPhrase": { "description": "Represents an example that the agent is trained on.", "id": "GoogleCloudDialogflowV2beta1IntentTrainingPhrase", @@ -14062,6 +14224,7 @@ "ARTICLE_SUGGESTION", "FAQ", "SMART_REPLY", + "DIALOGFLOW_ASSIST", "CONVERSATION_SUMMARIZATION" ], "enumDescriptions": [ @@ -14069,6 +14232,7 @@ "Run article suggestion model for chat.", "Run FAQ model.", "Run smart reply model for chat.", + "Run Dialogflow assist model for chat, which will return automated agent response as suggestion.", "Run conversation summarization model for chat." ], "type": "string" @@ -14119,6 +14283,29 @@ }, "type": "object" }, + "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse": { + "description": "The response message for Participants.SuggestDialogflowAssists.", + "id": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "properties": { + "contextSize": { + "description": "Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestDialogflowAssistsRequest.context_size field in the request if there aren't that many messages in the conversation.", + "format": "int32", + "type": "integer" + }, + "dialogflowAssistAnswers": { + "description": "Output only. Multiple reply options provided by Dialogflow assist service. The order is based on the rank of the model prediction.", + "items": { + "$ref": "GoogleCloudDialogflowV2beta1DialogflowAssistAnswer" + }, + "type": "array" + }, + "latestMessage": { + "description": "The name of the latest conversation message used to suggest answer. Format: `projects//locations//conversations//messages/`.", + "type": "string" + } + }, + "type": "object" + }, "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse": { "description": "The request message for Participants.SuggestFaqAnswers.", "id": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", @@ -14177,6 +14364,10 @@ "$ref": "GoogleCloudDialogflowV2beta1SuggestArticlesResponse", "description": "SuggestArticlesResponse if request is for ARTICLE_SUGGESTION." }, + "suggestDialogflowAssistsResponse": { + "$ref": "GoogleCloudDialogflowV2beta1SuggestDialogflowAssistsResponse", + "description": "SuggestDialogflowAssistsResponse if request is for DIALOGFLOW_ASSIST." + }, "suggestFaqAnswersResponse": { "$ref": "GoogleCloudDialogflowV2beta1SuggestFaqAnswersResponse", "description": "SuggestFaqAnswersResponse if request is for FAQ_ANSWER." @@ -14408,6 +14599,16 @@ "description": "Whether turn resulted in End Session page.", "type": "boolean" }, + "sentimentMagnitude": { + "description": "Sentiment magnitude of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, + "sentimentScore": { + "description": "Sentiment score of the user utterance if [sentiment](https://cloud.google.com/dialogflow/cx/docs/concept/sentiment) was enabled.", + "format": "float", + "type": "number" + }, "triggeredAbandonmentEvent": { "description": "Whether agent has triggered the event corresponding to user abandoning the conversation.", "type": "boolean" @@ -14456,7 +14657,7 @@ "type": "object" }, "GoogleCloudLocationLocation": { - "description": "A resource that represents Google Cloud Platform location.", + "description": "A resource that represents a Google Cloud location.", "id": "GoogleCloudLocationLocation", "properties": { "displayName": {