Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not using my voice input to create audio response #70

Open
Davidgomezrob opened this issue Nov 13, 2024 · 1 comment
Open

Not using my voice input to create audio response #70

Davidgomezrob opened this issue Nov 13, 2024 · 1 comment

Comments

@Davidgomezrob
Copy link

The assistant should use my microphone input (audio) to process responses dynamically. However, despite sending my audio data to the API, the model doesn’t seem to recognize or respond based on my actual audio input. Instead, it produces default responses, ignoring the context of my input audio.

Technical Setup:

  • Using Node.js with the realtime-api-beta package
  • Audio input is captured from the microphone, converted to PCM16 format, and streamed to the API.
  • The appendInputAudio() method is used to send audio chunks, followed by createResponse() to initiate a response when silence is detected.

Any insights on debugging this setup or ensuring the model correctly interprets the live audio input would be greatly appreciated!

@aarizirf
Copy link

aarizirf commented Dec 3, 2024

having the same issue ... any luck?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants