-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support audio outputs with openai provider #3907
Labels
Comments
Can you provide details on which information you would need to be returned? currently only audio inputs are supported |
Ideally, ai sdk should formally support audio, as it supports images and pdf. Other ai providers may as well implement audio i/o soon enough, i.e. anthropic & hume ai colaboration rumours are around. |
lgrammel
changed the title
When using sdk/openai, important data is discarded by the SDK
Support audio outputs with openai provider
Dec 3, 2024
lgrammel
added
enhancement
New feature or request
and removed
bug
Something isn't working
labels
Dec 3, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Description
When I use the
gpt-4o-audio-preview
model and pass the audio, the fetch method I've implemented myself, chat withopenai
to require the audio and transcription, the audio in the returned data was discarded, rendering the entire conversation completely invalid.Code example
No response
AI provider
@ai-sdk/openai v1.0.4
Additional context
I found that the zodSchema.safeParse method dropped the audio field.
The text was updated successfully, but these errors were encountered: