Dialogflow V2 API - Class Google::Cloud::Dialogflow::V2::StreamingDetectIntentRequest (v1.1.1)

Reference documentation and code samples for the Dialogflow V2 API class Google::Cloud::Dialogflow::V2::StreamingDetectIntentRequest.

The top-level message sent by the client to the Sessions.StreamingDetectIntent method.

Multiple request messages should be sent in order:

  1. The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
  2. If query_input was set to query_input.audio_config, all subsequent messages must contain input_audio to continue with Speech recognition. If you decide to rather detect an intent from text input after you already started Speech recognition, please send a message with query_input.text.

    However, note that:

* Dialogflow will bill you for the audio duration so far.
* Dialogflow discards all Speech recognition results in favor of the
  input text.
* Dialogflow will use the language code from the first message.

After you sent all input, you must half-close or abort the request stream.

Inherits

  • Object

Extended By

  • Google::Protobuf::MessageExts::ClassMethods

Includes

  • Google::Protobuf::MessageExts

Methods

#enable_debugging_info

def enable_debugging_info() -> ::Boolean
Returns
  • (::Boolean) — if true, StreamingDetectIntentResponse.debugging_info will get populated.

#enable_debugging_info=

def enable_debugging_info=(value) -> ::Boolean
Parameter
  • value (::Boolean) — if true, StreamingDetectIntentResponse.debugging_info will get populated.
Returns
  • (::Boolean) — if true, StreamingDetectIntentResponse.debugging_info will get populated.

#input_audio

def input_audio() -> ::String
Returns
  • (::String) — The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

#input_audio=

def input_audio=(value) -> ::String
Parameter
  • value (::String) — The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.
Returns
  • (::String) — The input audio content to be recognized. Must be sent if query_input was set to a streaming input audio config. The complete audio over all streaming messages must not exceed 1 minute.

#output_audio_config

def output_audio_config() -> ::Google::Cloud::Dialogflow::V2::OutputAudioConfig
Returns

#output_audio_config=

def output_audio_config=(value) -> ::Google::Cloud::Dialogflow::V2::OutputAudioConfig
Parameter
Returns

#output_audio_config_mask

def output_audio_config_mask() -> ::Google::Protobuf::FieldMask
Returns

#output_audio_config_mask=

def output_audio_config_mask=(value) -> ::Google::Protobuf::FieldMask
Parameter
Returns

#query_input

def query_input() -> ::Google::Cloud::Dialogflow::V2::QueryInput
Returns
  • (::Google::Cloud::Dialogflow::V2::QueryInput) —

    Required. The input specification. It can be set to:

    1. an audio config which instructs the speech recognizer how to process the speech audio,

    2. a conversational query in the form of text, or

    3. an event that specifies which intent to trigger.

#query_input=

def query_input=(value) -> ::Google::Cloud::Dialogflow::V2::QueryInput
Parameter
  • value (::Google::Cloud::Dialogflow::V2::QueryInput) —

    Required. The input specification. It can be set to:

    1. an audio config which instructs the speech recognizer how to process the speech audio,

    2. a conversational query in the form of text, or

    3. an event that specifies which intent to trigger.

Returns
  • (::Google::Cloud::Dialogflow::V2::QueryInput) —

    Required. The input specification. It can be set to:

    1. an audio config which instructs the speech recognizer how to process the speech audio,

    2. a conversational query in the form of text, or

    3. an event that specifies which intent to trigger.

#query_params

def query_params() -> ::Google::Cloud::Dialogflow::V2::QueryParameters
Returns

#query_params=

def query_params=(value) -> ::Google::Cloud::Dialogflow::V2::QueryParameters
Parameter
Returns

#session

def session() -> ::String
Returns
  • (::String) — Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters.

    For more information, see the API interactions guide.

    Note: Always use agent versions for production traffic. See Versions and environments.

#session=

def session=(value) -> ::String
Parameter
  • value (::String) — Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters.

    For more information, see the API interactions guide.

    Note: Always use agent versions for production traffic. See Versions and environments.

Returns
  • (::String) — Required. The name of the session the query is sent to. Format of the session name: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment. If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters.

    For more information, see the API interactions guide.

    Note: Always use agent versions for production traffic. See Versions and environments.

#single_utterance

def single_utterance() -> ::Boolean
Returns
  • (::Boolean) — Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.

#single_utterance=

def single_utterance=(value) -> ::Boolean
Parameter
  • value (::Boolean) — Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.
Returns
  • (::Boolean) — Please use InputAudioConfig.single_utterance instead. If false (default), recognition does not cease until the client closes the stream. If true, the recognizer will detect a single spoken utterance in input audio. Recognition ceases when it detects the audio's voice has stopped or paused. In this case, once a detected intent is received, the client should close the stream and start a new request with a new stream as needed. This setting is ignored when query_input is a piece of text or an event.