Processes a natural language query and returns structured, actionable data as a result. This method is not idempotent, because it may cause contexts and session entity types to be updated, which in turn might affect results of future queries. If you might use Agent Assist or other CCAI products now or in the future, consider using AnalyzeContent instead of DetectIntent
. AnalyzeContent
has additional functionality for Agent Assist and other CCAI products. Note: Always use agent versions for production traffic. See Versions and environments.
Scopes
You will need authorization for at least one of the following scopes to make a valid call:
- https://www.googleapis.com/auth/cloud-platform
- https://www.googleapis.com/auth/dialogflow
If unset, the scope for this method defaults to https://www.googleapis.com/auth/cloud-platform.
You can set the scope for this method like this: dialogflow2 --scope <scope> projects locations-agent-environments-users-sessions-detect-intent ...
Required Scalar Argument
- <session> (string)
- Required. The name of the session this query is sent to. Format:
projects//agent/sessions/
, orprojects//agent/environments//users//sessions/
. IfEnvironment ID
is not specified, we assume default 'draft' environment (Environment ID
might be referred to as environment name at some places). IfUser ID
is not specified, we are using "-". It's up to the API caller to choose an appropriateSession ID
andUser Id
. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession ID
andUser ID
must not exceed 36 characters. For more information, see the API interactions guide. Note: Always use agent versions for production traffic. See Versions and environments.
- Required. The name of the session this query is sent to. Format:
Required Request Value
The request value is a data-structure with various fields. Each field may be a simple scalar or another data-structure. In the latter case it is advised to set the field-cursor to the data-structure's field to specify values more concisely.
For example, a structure like this:
GoogleCloudDialogflowV2DetectIntentRequest:
input-audio: string
output-audio-config:
audio-encoding: string
sample-rate-hertz: integer
synthesize-speech-config:
effects-profile-id: [string]
pitch: number
speaking-rate: number
voice:
name: string
ssml-gender: string
volume-gain-db: number
output-audio-config-mask: string
query-input:
audio-config:
audio-encoding: string
disable-no-speech-recognized-event: boolean
enable-automatic-punctuation: boolean
enable-word-info: boolean
language-code: string
model: string
model-variant: string
opt-out-conformer-model-migration: boolean
phrase-hints: [string]
sample-rate-hertz: integer
single-utterance: boolean
event:
language-code: string
name: string
text:
language-code: string
text: string
query-params:
geo-location:
latitude: number
longitude: number
platform: string
reset-contexts: boolean
sentiment-analysis-request-config:
analyze-query-text-sentiment: boolean
time-zone: string
webhook-headers: { string: string }
can be set completely with the following arguments which are assumed to be executed in the given order. Note how the cursor position is adjusted to the respective structures, allowing simple field names to be used most of the time.
-r . input-audio=ea
- The natural language speech audio to be processed. This field should be populated iff
query_input
is set to an input audio config. A single request can contain up to 1 minute of speech audio data.
- The natural language speech audio to be processed. This field should be populated iff
output-audio-config audio-encoding=sea
- Required. Audio encoding of the synthesized audio content.
sample-rate-hertz=67
- The synthesis sample rate (in hertz) for this audio. If not provided, then the synthesizer will use the default sample rate based on the audio encoding. If this is different from the voice's natural sample rate, then the synthesizer will honor this request by converting to the desired sample rate (which might result in worse audio quality).
synthesize-speech-config effects-profile-id=dolores
- Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
- Each invocation of this argument appends the given value to the array.
pitch=0.10599511191550304
- Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
speaking-rate=0.6223860992915429
- Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
voice name=et
- Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender.
-
ssml-gender=est
- Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
-
.. volume-gain-db=0.6784614280826983
- Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
-
... output-audio-config-mask=et
- Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level. If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
query-input.audio-config audio-encoding=labore
- Required. Audio encoding of the audio content to process.
disable-no-speech-recognized-event=false
- Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If
false
and recognition doesn't return any result, triggerNO_SPEECH_RECOGNIZED
event to Dialogflow agent.
- Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If
enable-automatic-punctuation=false
- Enable automatic punctuation option at the speech backend.
enable-word-info=true
- If
true
, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information.
- If
language-code=gubergren
- Required. The language of the supplied audio. Dialogflow does not do translations. See Language Support for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language.
model=est
- Optional. Which Speech model to select for the given request. For more information, see Speech models.
model-variant=rebum.
- Which variant of the Speech model to use.
opt-out-conformer-model-migration=true
- If
true
, the request will opt out for STT conformer model migration. This field will be deprecated once force migration takes place in June 2024. Please refer to Dialogflow ES Speech model migration.
- If
phrase-hints=erat
- A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details. This field is deprecated. Please use
speech_contexts
instead. If you specify bothphrase_hints
andspeech_contexts
, Dialogflow will treat thephrase_hints
as a single additionalSpeechContext
. - Each invocation of this argument appends the given value to the array.
- A list of strings containing words and phrases that the speech recognizer should recognize with higher likelihood. See the Cloud Speech documentation for more details. This field is deprecated. Please use
sample-rate-hertz=2
- Required. Sample rate (in Hertz) of the audio content sent in the query. Refer to Cloud Speech API documentation for more details.
-
single-utterance=true
- If
false
(default), recognition does not cease until the client closes the stream. Iftrue
, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. Note: This setting is relevant only for streaming methods. Note: When specified, InputAudioConfig.single_utterance takes precedence over StreamingDetectIntentRequest.single_utterance.
- If
-
..event language-code=eirmod
- Required. The language of this query. See Language Support for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. This field is ignored when used in the context of a WebhookResponse.followup_event_input field, because the language was already defined in the originating detect intent request.
-
name=amet.
- Required. The unique identifier of the event.
-
..text language-code=takimata
- Required. The language of this conversational query. See Language Support for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language.
-
text=amet.
- Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters for virtual agent interactions.
-
...query-params.geo-location latitude=0.6237467640152612
- The latitude in degrees. It must be in the range [-90.0, +90.0].
-
longitude=0.4577859622777416
- The longitude in degrees. It must be in the range [-180.0, +180.0].
-
.. platform=sit
- The platform of the virtual agent response messages. If not empty, only emits messages from this platform in the response. Valid values are the enum names of platform.
reset-contexts=false
- Specifies whether to delete all contexts in the current session before the new ones are activated.
-
sentiment-analysis-request-config analyze-query-text-sentiment=false
- Instructs the service to perform sentiment analysis on
query_text
. If not provided, sentiment analysis is not performed onquery_text
.
- Instructs the service to perform sentiment analysis on
-
.. time-zone=voluptua.
- The time zone of this conversational query from the time zone database, e.g., America/New_York, Europe/Paris. If not provided, the time zone specified in agent settings is used.
webhook-headers=key=justo
- This field can be used to pass HTTP headers for a webhook call. These headers will be sent to webhook along with the headers that have been configured through the Dialogflow web console. The headers defined within this field will overwrite the headers configured through the Dialogflow console if there is a conflict. Header names are case-insensitive. Google's specified headers are not allowed. Including: "Host", "Content-Length", "Connection", "From", "User-Agent", "Accept-Encoding", "If-Modified-Since", "If-None-Match", "X-Forwarded-For", etc.
- the value will be associated with the given
key
About Cursors
The cursor position is key to comfortably set complex nested structures. The following rules apply:
- The cursor position is always set relative to the current one, unless the field name starts with the
.
character. Fields can be nested such as in-r f.s.o
. - The cursor position is set relative to the top-level structure if it starts with
.
, e.g.-r .s.s
- You can also set nested fields without setting the cursor explicitly. For example, to set a value relative to the current cursor position, you would specify
-r struct.sub_struct=bar
. - You can move the cursor one level up by using
..
. Each additional.
moves it up one additional level. E.g....
would go three levels up.
Optional Output Flags
The method's return value a JSON encoded structure, which will be written to standard output by default.
- -o out
- out specifies the destination to which to write the server's result to.
It will be a JSON-encoded structure.
The destination may be
-
to indicate standard output, or a filepath that is to contain the received bytes. If unset, it defaults to standard output.
- out specifies the destination to which to write the server's result to.
It will be a JSON-encoded structure.
The destination may be
Optional General Properties
The following properties can configure any call, and are not specific to this method.
-
-p $-xgafv=string
- V1 error format.
-
-p access-token=string
- OAuth access token.
-
-p alt=string
- Data format for response.
-
-p callback=string
- JSONP
-
-p fields=string
- Selector specifying which fields to include in a partial response.
-
-p key=string
- API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.
-
-p oauth-token=string
- OAuth 2.0 token for the current user.
-
-p pretty-print=boolean
- Returns response with indentations and line breaks.
-
-p quota-user=string
- Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.
-
-p upload-type=string
- Legacy upload protocol for media (e.g. "media", "multipart").
-
-p upload-protocol=string
- Upload protocol for media (e.g. "raw", "multipart").