TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

OpenAI DevDay 2024 live blog

212 点作者 plurby7 个月前

20 条评论

qwertox7 个月前
&gt; The Realtime API improves this by streaming audio inputs and outputs directly, enabling more natural conversational experiences. It can also handle interruptions automatically, much like Advanced Voice Mode in ChatGPT.<p>&gt; Under the hood, the Realtime API lets you create a persistent WebSocket connection to exchange messages with GPT-4o. The API supports function calling(opens in a new window), which makes it possible for voice assistants to respond to user requests by triggering actions or pulling in new context.<p>-<p>This sounds really interesting, and I see a great use cases for it. However, I&#x27;m wondering if the API provides a text transcription of both the input and output so that I can store the data directly in a database without needing to transcribe the audio separately.<p>-<p>Edit: Apparently it does.<p>It sends `conversation.item.input_audio_transcription.completed` [0] events when the input transcription is done (I guess a couple of them in real-time)<p>and `response.done` [1] with the response text.<p>[0] <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;api-reference&#x2F;realtime-server-events&#x2F;conversation-item-input-audio-transcription-completed" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;api-reference&#x2F;realtime-serv...</a><p>[1] <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;api-reference&#x2F;realtime-server-events&#x2F;response-done" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;api-reference&#x2F;realtime-serv...</a>
评论 #41714081 未加载
评论 #41715209 未加载
评论 #41713202 未加载
siva77 个月前
I&#x27;ve never seen a company publishing consistently groundbreaking features at such a speed like this one. I really wonder how their teams work. It&#x27;s unprecedented at what i&#x27;ve seen in 15 years software
评论 #41712839 未加载
评论 #41713332 未加载
评论 #41712873 未加载
评论 #41715865 未加载
评论 #41714637 未加载
ponty_rick7 个月前
&gt; 11:43 Fields are generated in the same order that you defined them in the schema, even though JSON is supposed to ignore key order. This ensures you can implement things like chain-of-thought by adding those keys in the correct order in your schema design.<p>Why not use an array of key value pairs if you want to maintain ordering without breaking traditional JSON rules?<p>[ {key1:value1}, {key2:value2} ]
评论 #41713971 未加载
评论 #41713383 未加载
serjester7 个月前
The eval platform is a game changer.<p>It&#x27;s nice to have have a solution from OpenAI given how much they use a variant of this internally. I&#x27;ve tried like 5 YC startups and I don&#x27;t think anyone&#x27;s really solved this.<p>There&#x27;s the very real risk of vendor lock-in but quickly scanning the docs seems like it&#x27;s a pretty portable implementation.
alach117 个月前
It&#x27;s pretty amazing that they made prompt caching automatic. It&#x27;s rare that a company gives a 50% discount without the customer explicitly requesting it! Of course... they might be retaining some margin, judging by their discount being 50% vs. Anthropic&#x27;s 90%.
评论 #41714519 未加载
thenameless77417 个月前
Blog updates:<p>- Introducing the Realtime API: <a href="https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-the-realtime-api&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-the-realtime-api&#x2F;</a><p>- Introducing vision to the fine-tuning API: <a href="https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-vision-to-the-fine-tuning-api&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-vision-to-the-fine-tuni...</a><p>- Prompt Caching in the API: <a href="https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;api-prompt-caching&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;api-prompt-caching&#x2F;</a><p>- Model Distillation in the API: <a href="https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;api-model-distillation&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;api-model-distillation&#x2F;</a><p>Docs updates:<p>- Realtime API: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;realtime" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;realtime</a><p>- Vision fine-tuning: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;fine-tuning&#x2F;vision" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;fine-tuning&#x2F;vision</a><p>- Prompt Caching: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;prompt-caching" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;prompt-caching</a><p>- Model Distillation: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;distillation" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;distillation</a><p>- Evaluating model performance: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;evals" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;docs&#x2F;guides&#x2F;evals</a><p>Additional updates from @OpenAIDevs: <a href="https:&#x2F;&#x2F;x.com&#x2F;OpenAIDevs&#x2F;status&#x2F;1841175537060102396" rel="nofollow">https:&#x2F;&#x2F;x.com&#x2F;OpenAIDevs&#x2F;status&#x2F;1841175537060102396</a><p>- New prompt generator on <a href="https:&#x2F;&#x2F;playground.openai.com" rel="nofollow">https:&#x2F;&#x2F;playground.openai.com</a><p>- Access to the o1 model is expanded to developers on usage tier 3, and rate limits are increased (to the same limits as GPT-4o)<p>Additional updates from @OpenAI: <a href="https:&#x2F;&#x2F;x.com&#x2F;OpenAI&#x2F;status&#x2F;1841179938642411582" rel="nofollow">https:&#x2F;&#x2F;x.com&#x2F;OpenAI&#x2F;status&#x2F;1841179938642411582</a><p>- Advanced Voice is rolling out globally to ChatGPT Enterprise, Edu, and Team users. Free users will get a sneak peak of it (except EU).
评论 #41712419 未加载
1010087 个月前
I understand the Realtime API voice novelty, and the techonological achievement it is, but I don&#x27;t see it from the product point of view. It looks like one of those startups finding a solution before knowing the problem.<p>The two examples shown in the DevDay are the things I don&#x27;t really want to do in the future. I don&#x27;t want to talk to anybody, and I don&#x27;t want to wait for their answer in a human form. That&#x27;s why I order my food through an app or Whatsapp, or why I prefer to buy my tickets online. In the rare case I call to order food, it&#x27;s because I have a weird question or a weird request (can I pick it up in X minutes? Can you prepare it in a different way?)<p>I hope we don&#x27;t start seeing apps using conversations as interfaces because it would really horrible (leaving aside the fact that a lot of people don&#x27;t know how to communicate themselves, different accents, sound environments, etc), while clicking or typing work almost the same for everyone (at least much more normalized than talking)
评论 #41713998 未加载
评论 #41715887 未加载
评论 #41715176 未加载
评论 #41713597 未加载
评论 #41714101 未加载
superdisk7 个月前
Holy crud, I figured they would guard this for a long time and I was really salivating to make some stuff with it. The doors are wide open for all sorts of stuff now, Advanced Voice is the first feature since ChatGPT initially came out that really has my jaw on the floor.
评论 #41712600 未加载
minimaxir7 个月前
From the Realtime API blog post: <a href="https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-the-realtime-api&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;index&#x2F;introducing-the-realtime-api&#x2F;</a><p>&gt; Audio in the Chat Completions API will be released in the coming weeks, as a new model `gpt-4o-audio-preview`. With `gpt-4o-audio-preview`, developers can input text or audio into GPT-4o and receive responses in text, audio, or both.<p>&gt; The Realtime API uses both text tokens and audio tokens. Text input tokens are priced at $5 per 1M and $20 per 1M output tokens. Audio input is priced at $100 per 1M tokens and output is $200 per 1M tokens. This equates to approximately $0.06 per minute of audio input and $0.24 per minute of audio output. Audio in the Chat Completions API will be the same price.<p>As usual, OpenAI failed to emphasize the real-game changer feature at their Dev Day: audio output from the standard generation API.<p>This has severe implications for text-to-speech apps, particularly if the audio output style is as steerable as the gpt-4o voice demos.
评论 #41712910 未加载
N_A_T_E7 个月前
I just need their API to be faster. 15-30 seconds per request using 4o-mini isn&#x27;t good enough for responsive applications.
评论 #41714363 未加载
评论 #41715298 未加载
评论 #41714076 未加载
评论 #41715116 未加载
simonw7 个月前
For anyone who’s interested, I’ve written up details of how the underlying live blog system works here: <a href="https:&#x2F;&#x2F;til.simonwillison.net&#x2F;django&#x2F;live-blog" rel="nofollow">https:&#x2F;&#x2F;til.simonwillison.net&#x2F;django&#x2F;live-blog</a>
modeless7 个月前
I didn&#x27;t expect an API for advanced voice so soon. That&#x27;s pretty great. Here&#x27;s the thing I was really wondering: Audio is $.06&#x2F;min in, $.24&#x2F;min out. Can&#x27;t wait to try some language learning apps built with this. It&#x27;ll also be fun for controlling robots.
sammyteee7 个月前
Loving these live updates, keep em coming! Thanks Simon!
nielsole7 个月前
&gt; The first big announcement: a realtime API, providing the ability to use WebSockets to implement voice input and output against their models.<p>I guess this is using their &quot;old&quot; turn-based voice system?
评论 #41712262 未加载
cedws7 个月前
WebSockets for realtime? WS is TCP based, wouldn’t it be better to use something UDP based if you want to optimise for latency?
og_kalu7 个月前
Image output for 4o in the API would be very nice but i&#x27;m not sure if that&#x27;s at all in the cards.<p>Audio output in the api now but you lose image input. Why ? That&#x27;s a shame.
jbaudanza7 个月前
Interesting choice of a 24kHz sample rate for PCM audio. I wonder if the model was trained on 24kHz audio, rather than the usual 8&#x2F;16kHz for ML models.
hidelooktropic7 个月前
Any word on increased weekly caps on o1 usage?
评论 #41714335 未加载
lysecret7 个月前
Using structured outputs for generative ui is such a cool idea does anyone know some cool web demos related to this ?
评论 #41714598 未加载
bigcat123456787 个月前
Seems mostly standard items so far.