You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The transcription is expected to be empty when the conversation item is created. The transcription should be included in a message sent later by the realtime API, and the user_speech_committed event will be emitted when the agent receives the transcription.
There should be a debug log for committed user speech for example
Hmm, I restarted my docker container, without having changed anything, and now the event is being picked up it looks like, and I'm seeing events trigger on both sides now, sorry for the errant issue.
I will say though that the transcription is radically different from actual audio that the AI picked up and used. I'm imagining this is due to discrepancies between the realtime model and the whisper model used to generate the transcript?
I'm not sure if there's anything I can do to improve that, though.
Hello - I'm not sure if this is a bug, or just something I'm doing wrong.
I am creating a model:
I have event handlers defined for when speech is committed:
During a conversation, the
agent_speech_committed
event is fired normally and themsg
param contains the AI's response.However, the
user_speech_committed
event is never picked up.In addition, in the debug logs, I can see a user conversation item being created with audio, but the transcription is blank:
I'm not sure if that could be related to the event not firing or not.
The text was updated successfully, but these errors were encountered: