You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/AssistantEnd.ts
+5-1
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,11 @@
6
6
* When provided, the output is an assistant end message.
7
7
*/
8
8
exportinterfaceAssistantEnd{
9
-
/** The type of message sent through the socket; for an Assistant End message, this must be `assistant_end`. */
9
+
/**
10
+
* The type of message sent through the socket; for an Assistant End message, this must be `assistant_end`.
11
+
*
12
+
* This message indicates the conclusion of the assistant’s response, signaling that the assistant has finished speaking for the current conversational turn.
13
+
*/
10
14
type: "assistant_end";
11
15
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/AssistantInput.ts
+6-2
Original file line number
Diff line number
Diff line change
@@ -6,10 +6,14 @@
6
6
* When provided, the input is spoken by EVI.
7
7
*/
8
8
exportinterfaceAssistantInput{
9
-
/** The type of message sent through the socket; for an Assistant Input message, this must be `assistant_input`. */
9
+
/** The type of message sent through the socket; must be `assistant_input` for our server to correctly identify and process it as an Assistant Input message. */
10
10
type: "assistant_input";
11
11
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
12
12
customSessionId?: string;
13
-
/** Text to be synthesized. */
13
+
/**
14
+
* Assistant text to synthesize into spoken audio and insert into the conversation.
15
+
*
16
+
* EVI uses this text to generate spoken audio using our proprietary expressive text-to-speech model. Our model adds appropriate emotional inflections and tones to the text based on the user’s expressions and the context of the conversation. The synthesized audio is streamed back to the user as an [Assistant Message](/reference/empathic-voice-interface-evi/chat/chat#receive.Assistant%20Message.type).
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/AssistantMessage.ts
+7-3
Original file line number
Diff line number
Diff line change
@@ -8,16 +8,20 @@ import * as Hume from "../../../index";
8
8
* When provided, the output is an assistant message.
9
9
*/
10
10
exportinterfaceAssistantMessage{
11
-
/** The type of message sent through the socket; for an Assistant Message, this must be `assistant_message`. */
11
+
/**
12
+
* The type of message sent through the socket; for an Assistant Message, this must be `assistant_message`.
13
+
*
14
+
* This message contains both a transcript of the assistant’s response and the expression measurement predictions of the assistant’s audio output.
15
+
*/
12
16
type: "assistant_message";
13
17
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
14
18
customSessionId?: string;
15
-
/** ID of the assistant message. */
19
+
/** ID of the assistant message. Allows the Assistant Message to be tracked and referenced. */
16
20
id?: string;
17
21
/** Transcript of the message. */
18
22
message: Hume.empathicVoice.ChatMessage;
19
23
/** Inference model results. */
20
24
models: Hume.empathicVoice.Inference;
21
-
/** Indicates if this message was constructed from a text input message. */
25
+
/** Indicates if this message was inserted into the conversation as text from an [Assistant Input message](/reference/empathic-voice-interface-evi/chat/chat#send.Assistant%20Input.text). */
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/AudioInput.ts
+12-2
Original file line number
Diff line number
Diff line change
@@ -6,10 +6,20 @@
6
6
* When provided, the input is audio.
7
7
*/
8
8
exportinterfaceAudioInput{
9
-
/** The type of message sent through the socket; for an Audio Input message, this must be `audio_input`. */
9
+
/**
10
+
* The type of message sent through the socket; must be `audio_input` for our server to correctly identify and process it as an Audio Input message.
11
+
*
12
+
* This message is used for sending audio input data to EVI for processing and expression measurement. Audio data should be sent as a continuous stream, encoded in Base64.
13
+
*/
10
14
type: "audio_input";
11
15
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
12
16
customSessionId?: string;
13
-
/** Base64 encoded audio input. */
17
+
/**
18
+
* Base64 encoded audio input to insert into the conversation.
19
+
*
20
+
* The audio input must be captured and transmitted to EVI as a continuous stream, with the audio data sent in small chunks for better transcription quality. When capturing audio through the browser, we recommend recording the audio in 100ms intervals and adjusting from there to determine if smaller or larger chunks are needed. These chunks should be continuously sent to EVI as Audio Input messages.
21
+
*
22
+
* The content of an Audio Input message is treated as the user’s speech to EVI. EVI processes the audio, conducts expression measurement using the prosody model, and responds accordingly.
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/AudioOutput.ts
+2-2
Original file line number
Diff line number
Diff line change
@@ -10,8 +10,8 @@ export interface AudioOutput {
10
10
type: "audio_output";
11
11
/** Used to manage conversational state, correlate frontend and backend data, and persist conversations across EVI sessions. */
12
12
customSessionId?: string;
13
-
/** ID of the audio output. */
13
+
/** ID of the audio output. Allows the Audio Output message to be tracked and referenced. */
14
14
id: string;
15
-
/** Base64 encoded audio output. */
15
+
/** Base64 encoded audio output. This encoded audio is transmitted to the client, where it can be decoded and played back as part of the user interaction. */
Copy file name to clipboardExpand all lines: src/api/resources/empathicVoice/types/BuiltinToolConfig.ts
+1
Original file line number
Diff line number
Diff line change
@@ -6,5 +6,6 @@ import * as Hume from "../../../index";
6
6
7
7
exportinterfaceBuiltinToolConfig{
8
8
name: Hume.empathicVoice.BuiltInTool;
9
+
/** Optional text passed to the supplemental LLM if the tool call fails. The LLM then uses this text to generate a response back to the user, ensuring continuity in the conversation. */
0 commit comments