Story Teller
This guide will help you to run the Story Teller usecase with Large Language Model in the TEN-Agent Playground.
STT + TTS + LLM
Prerequisites
Make sure you have the TEN-Agent playground running. If not, follow the Run Playground guide to start the playground.
You will need following information prepared:
STT info, any supported STT can be used. Deepgram is relatively easy to register and get started with.
TTS info, any supported TTS can be used. Fish.Audio is relatively easy to register and get started with.
LLM info, For this use case only OpenAI or OpenAI API compatible models are supported.
RTC info, currently only Agora RTC is supported. You can register your account at Agora. We assume you have your App ID and App Certificate ready when you configure your
.env
file.
Steps
Open the playground at localhost:3000 to configure your agent.
Select the graph type
story_teller
.Click on
Module Picker
to open the module selection.If you preferred STT/TTS module is not by default selected, you can select the module from the dropdown list. Note you will need to configure the module with the correct information like API key, etc.
The
LLM
module is preconfigured to haveOpenAI ChatGPT
selected, don't change it.Click on
Save Change
to apply the module to the graph.Click on the Button to the right of the graph selection to open the property configuration. You will see a list of properties that can be configured for the selected Large Language Model.
Configure the properties with the information you prepared.
Click on
Save Change
to apply the properties to the Large Language Model.If you see the success toast, the properties are successfully applied to the Large Language Model.
You are all set! Now you can start speaking to the Voice Assistant by clicking on the
Connect
button. Note you will need to wait a few seconds for agent to initialzie itself.
Using Azure STT
Azure STT is integrated within RTC extension module. That's why if you want to use Azure STT, you will need to select story_teller_integrated_stt
graph type.
Bind Tools
The story_teller use case is preconfigured to use openai_image_generate_tool
, so usually you don't need to change anything.
Realtime V2V
Prerequisites
Make sure you have the TEN-Agent playground running. If not, follow the Run Playground guide to start the playground.
You will need following information prepared:
Realtime API Key
RTC info, currently only Agora RTC is supported. You can register your account at Agora. We assume you have your App ID and App Certificate ready when you configure your
.env
file.
Steps
Open the playground at localhost:3000 to configure your agent.
Select the graph type
story_teller_realtime
.Click on
Module Picker
to open the module selection.The
V2V
module is preconfigured to haveOpenAI Realtime
selected. You can select other V2V modules from the dropdown list if needed. Note you will need to copy theprompt
property from theOpenAI Realtime
module to the new module, as module properties will be reset to default when switching.Click on
Save Change
to apply the module to the graph if you have changed the V2V module, or if you have not changed the V2V module, you can skip this step.Click on the Button to the right of the graph selection to open the property configuration. You will see a list of properties that can be configured for the selected V2V module.
Configure the
Realtime API Key
property with the information you prepared. If you have changed the V2V module in previous steps, do remember to copy theprompt
property from theOpenAI Realtime
module to the new module.Click on
Save Change
to apply the property to the V2V module.If you see the success toast, the property is successfully applied to the V2V module.
You are all set! Now you can start speaking to the Voice Assistant by clicking on the
Connect
button. Note you will need to wait a few seconds for agent to initialzie itself.
Bind Tools
The story_teller_realtime use case is preconfigured to use openai_image_generate_tool
, so usually you don't need to change anything.
Last updated