Generates and returns a summary for a conversation that does not have a resource created for it.

Scopes

You will need authorization for at least one of the following scopes to make a valid call:

  • https://www.googleapis.com/auth/cloud-platform
  • https://www.googleapis.com/auth/dialogflow

If unset, the scope for this method defaults to https://www.googleapis.com/auth/cloud-platform. You can set the scope for this method like this: dialogflow2-beta1 --scope <scope> projects locations-suggestions-generate-stateless-summary ...

Required Scalar Argument

  • <parent> (string)
    • Required. The parent resource to charge for the Summary's generation. Format: projects//locations/.

Required Request Value

The request value is a data-structure with various fields. Each field may be a simple scalar or another data-structure. In the latter case it is advised to set the field-cursor to the data-structure's field to specify values more concisely.

For example, a structure like this:

GoogleCloudDialogflowV2beta1GenerateStatelessSummaryRequest:
  conversation-profile:
    automated-agent-config:
      agent: string
      session-ttl: string
    create-time: string
    display-name: string
    human-agent-assistant-config:
      end-user-suggestion-config:
        group-suggestion-responses: boolean
      human-agent-suggestion-config:
        group-suggestion-responses: boolean
      message-analysis-config:
        enable-entity-extraction: boolean
        enable-sentiment-analysis: boolean
      notification-config:
        message-format: string
        topic: string
    human-agent-handoff-config:
      live-person-config:
        account-number: string
      salesforce-live-agent-config:
        button-id: string
        deployment-id: string
        endpoint-domain: string
        organization-id: string
    language-code: string
    logging-config:
      enable-stackdriver-logging: boolean
    name: string
    new-message-event-notification-config:
      message-format: string
      topic: string
    notification-config:
      message-format: string
      topic: string
    security-settings: string
    stt-config:
      model: string
      speech-model-variant: string
      use-timeout-based-endpointing: boolean
    time-zone: string
    tts-config:
      effects-profile-id: [string]
      pitch: number
      speaking-rate: number
      voice:
        name: string
        ssml-gender: string
      volume-gain-db: number
    update-time: string
  latest-message: string
  max-context-size: integer

can be set completely with the following arguments which are assumed to be executed in the given order. Note how the cursor position is adjusted to the respective structures, allowing simple field names to be used most of the time.

  • -r .conversation-profile.automated-agent-config agent=vero
    • Required. ID of the Dialogflow agent environment to use. This project needs to either be the same project as the conversation or you need to grant service-@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow API Service Agent role in this project. - For ES agents, use format: projects//locations//agent/environments/. If environment is not specified, the default draft environment is used. Refer to DetectIntentRequest for more details. - For CX agents, use format projects//locations//agents//environments/. If environment is not specified, the default draft environment is used.
  • session-ttl=diam

    • Optional. Configure lifetime of the Dialogflow session. By default, a Dialogflow CX session remains active and its data is stored for 30 minutes after the last request is sent for the session. This value should be no longer than 1 day.
  • .. create-time=takimata

    • Output only. Create time of the conversation profile.
  • display-name=voluptua.
    • Required. Human readable name for this profile. Max length 1024 bytes.
  • human-agent-assistant-config.end-user-suggestion-config group-suggestion-responses=false

    • If group_suggestion_responses is false, and there are multiple feature_configs in event based suggestion or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or StreamingAnalyzeContentResponse. If group_suggestion_responses set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
  • ..human-agent-suggestion-config group-suggestion-responses=true

    • If group_suggestion_responses is false, and there are multiple feature_configs in event based suggestion or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or StreamingAnalyzeContentResponse. If group_suggestion_responses set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse.
  • ..message-analysis-config enable-entity-extraction=false

    • Enable entity extraction in conversation messages on agent assist stage. If unspecified, defaults to false. Currently, this feature is not general available, please contact Google to get access.
  • enable-sentiment-analysis=false

    • Enable sentiment analysis in conversation messages on agent assist stage. If unspecified, defaults to false. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral: https://cloud.google.com/natural-language/docs/basics#sentiment_analysis For Participants.StreamingAnalyzeContent method, result will be in StreamingAnalyzeContentResponse.message.SentimentAnalysisResult. For Participants.AnalyzeContent method, result will be in AnalyzeContentResponse.message.SentimentAnalysisResult For Conversations.ListMessages method, result will be in ListMessagesResponse.messages.SentimentAnalysisResult If Pub/Sub notification is configured, result will be in ConversationEvent.new_message_payload.SentimentAnalysisResult.
  • ..notification-config message-format=dolor

    • Format of message.
  • topic=ea

    • Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant service-@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow Service Agent role in the topic project. For chat integration to receive notification, make sure API caller has been granted the Dialogflow Service Agent role for the topic. Format: projects//locations//topics/.
  • ...human-agent-handoff-config.live-person-config account-number=eirmod

    • Required. Account number of the LivePerson account to connect. This is the account number you input at the login page.
  • ..salesforce-live-agent-config button-id=et

    • Required. Live Agent chat button ID.
  • deployment-id=erat
    • Required. Live Agent deployment ID.
  • endpoint-domain=no
    • Required. Domain of the Live Agent endpoint for this agent. You can find the endpoint URL in the Live Agent settings page. For example if URL has the form https://d.la4-c2-phx.salesforceliveagent.com/..., you should fill in d.la4-c2-phx.salesforceliveagent.com.
  • organization-id=amet

    • Required. The organization ID of the Salesforce account.
  • ... language-code=vero

    • Language code for the conversation profile. If not specified, the language is en-US. Language at ConversationProfile should be set for all non en-us languages. This should be a BCP-47 language tag. Example: "en-US".
  • logging-config enable-stackdriver-logging=false

    • Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.
  • .. name=dolor

    • The unique identifier of this conversation profile. Format: projects//locations//conversationProfiles/.
  • new-message-event-notification-config message-format=est
    • Format of message.
  • topic=et

    • Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant service-@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow Service Agent role in the topic project. For chat integration to receive notification, make sure API caller has been granted the Dialogflow Service Agent role for the topic. Format: projects//locations//topics/.
  • ..notification-config message-format=ipsum

    • Format of message.
  • topic=stet

    • Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos. For telephony integration to receive notification, make sure either this topic is in the same project as the conversation or you grant service-@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow Service Agent role in the topic project. For chat integration to receive notification, make sure API caller has been granted the Dialogflow Service Agent role for the topic. Format: projects//locations//topics/.
  • .. security-settings=stet

    • Name of the CX SecuritySettings reference for the agent. Format: projects//locations//securitySettings/.
  • stt-config model=amet.
    • Which Speech model to select. Select the model best suited to your domain to get best results. If a model is not explicitly specified, then Dialogflow auto-selects a model based on other parameters in the SpeechToTextConfig and Agent settings. If enhanced speech model is enabled for the agent and an enhanced version of the specified model for the language does not exist, then the speech is recognized using the standard version of the specified model. Refer to Cloud Speech API documentation for more details. If you specify a model, the following models typically have the best performance: - phone_call (best for Agent Assist and telephony) - latest_short (best for Dialogflow non-telephony) - command_and_search Leave this field unspecified to use Agent Speech settings for model selection.
  • speech-model-variant=ut
    • The speech model used in speech to text. SPEECH_MODEL_VARIANT_UNSPECIFIED, USE_BEST_AVAILABLE will be treated as USE_ENHANCED. It can be overridden in AnalyzeContentRequest and StreamingAnalyzeContentRequest request. If enhanced model variant is specified and an enhanced version of the specified model for the language does not exist, then it would emit an error.
  • use-timeout-based-endpointing=false

    • Use timeout based endpointing, interpreting endpointer sensitivy as seconds of timeout value.
  • .. time-zone=sed

    • The time zone of this conversational profile from the time zone database, e.g., America/New_York, Europe/Paris. Defaults to America/New_York.
  • tts-config effects-profile-id=accusam
    • Optional. An identifier which selects 'audio effects' profiles that are applied on (post synthesized) text to speech. Effects are applied on top of each other in the order they are given.
    • Each invocation of this argument appends the given value to the array.
  • pitch=0.9633944848352897
    • Optional. Speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 semitones from the original pitch. -20 means decrease 20 semitones from the original pitch.
  • speaking-rate=0.2236339943161314
    • Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.
  • voice name=no
    • Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender. For the list of available voices, please refer to Supported voices and languages.
  • ssml-gender=lorem

    • Optional. The preferred gender of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and name. Note that this is only a preference, not requirement. If a voice of the appropriate gender is not available, the synthesizer should substitute a voice with a different gender rather than failing the request.
  • .. volume-gain-db=0.3641019933182008

    • Optional. Volume gain (in dB) of the normal native volume supported by the specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) will play at approximately half the amplitude of the normal native signal amplitude. A value of +6.0 (dB) will play at approximately twice the amplitude of the normal native signal amplitude. We strongly recommend not to exceed +10 (dB) as there's usually no effective increase in loudness for any value greater than that.
  • .. update-time=justo

    • Output only. Update time of the conversation profile.
  • .. latest-message=duo

    • The name of the latest conversation message used as context for generating a Summary. If empty, the latest message of the conversation will be used. The format is specific to the user and the names of the messages provided.
  • max-context-size=60
    • Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 500 and at most 1000.

About Cursors

The cursor position is key to comfortably set complex nested structures. The following rules apply:

  • The cursor position is always set relative to the current one, unless the field name starts with the . character. Fields can be nested such as in -r f.s.o .
  • The cursor position is set relative to the top-level structure if it starts with ., e.g. -r .s.s
  • You can also set nested fields without setting the cursor explicitly. For example, to set a value relative to the current cursor position, you would specify -r struct.sub_struct=bar.
  • You can move the cursor one level up by using ... Each additional . moves it up one additional level. E.g. ... would go three levels up.

Optional Output Flags

The method's return value a JSON encoded structure, which will be written to standard output by default.

  • -o out
    • out specifies the destination to which to write the server's result to. It will be a JSON-encoded structure. The destination may be - to indicate standard output, or a filepath that is to contain the received bytes. If unset, it defaults to standard output.

Optional General Properties

The following properties can configure any call, and are not specific to this method.

  • -p $-xgafv=string

    • V1 error format.
  • -p access-token=string

    • OAuth access token.
  • -p alt=string

    • Data format for response.
  • -p callback=string

    • JSONP
  • -p fields=string

    • Selector specifying which fields to include in a partial response.
  • -p key=string

    • API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.
  • -p oauth-token=string

    • OAuth 2.0 token for the current user.
  • -p pretty-print=boolean

    • Returns response with indentations and line breaks.
  • -p quota-user=string

    • Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters.
  • -p upload-type=string

    • Legacy upload protocol for media (e.g. "media", "multipart").
  • -p upload-protocol=string

    • Upload protocol for media (e.g. "raw", "multipart").