TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Model Distillation in the API

64 pointsby GavCo7 months ago

7 comments

simonw7 months ago
They announced this at DevDay at the beginning of October.<p>It&#x27;s effectively a layer of (well needed) sugar on top of their existing fine-tuning mechanism.<p>The challenge with fine-tuning is collecting a representative dataset to tune against. The tooling they added makes it easy for you to persist your prompts and responses within the OpenAI platform, and then later select those persisted pairs (that were created with e.g. GPT-4o) and use them to fine-tune a cheaper model (like GPT-4o mini) - such that the more expensive model is effectively &quot;teaching&quot; the cheaper model what to do.<p>You could do this before, but it was a LOT of work. The new &quot;distillation&quot; features make it easier.
评论 #42010445 未加载
behnamoh7 months ago
With each announcement, OpenAI kills yet another class of startups. I wonder if there are areas that OpenAI (and other AI companies) can&#x27;t enter because those seem to be the only viable startup ideas in the long-term.<p>Currently, OAI does all the following:<p>- offers flagship models<p>- offers lite models<p>- offers easy finetuning of their models<p>- offers structured output and guaranteed JSON output.<p>- offers parallel tool&#x2F;function calling which remains unmatched.<p>- has low API costs.<p>- offers a nice UI for their models<p>- offers Mac, iOS, Android, Windows app clients.<p>- offers image generation capabilities INTEGRATED with their language models.<p>- offers two-tier subscription plans for ordinary&#x2F;pro (team) users.<p>- offers custom GPTs which can be used by ordinary people to create GPT experiences tailored to specific tasks (no need to build a website on your own).<p>- allows users to easily share chats!! (it took Anthropic a long time to have this feature, and even now it&#x27;s not as good as OpenAI&#x27;s solution).<p>- offers prompt caching and task scheduling to further save costs.<p>- offers unrivaled voice-to-text models at different sizes (Whisper).<p>- offers text-to-voice models that feel much more natural than the competition.<p>- has outstanding documentation.<p>- sets the standard for API (all other companies have to follow their conventions, such as `messages`, `.choices[0].message.content`, etc.)<p>- has the most capable team to, idk, build AGI&#x2F;ASI...
评论 #42009778 未加载
评论 #42010329 未加载
评论 #42010341 未加载
评论 #42009785 未加载
评论 #42010105 未加载
评论 #42011483 未加载
评论 #42010428 未加载
GavCo7 months ago
From today&#x27;s ChatGPT search announcement: &quot;The search model is a fine-tuned version of GPT-4o, post-trained using novel synthetic data generation techniques, including distilling outputs from OpenAI o1-preview.&quot;
patelajay2857 months ago
We&#x27;ve been working on a Python framework where one of the use cases is easy distillation from larger models to smaller open-source models and smaller-closed source models (where you don&#x27;t have to still use &#x2F; pay for the closed-source API service): <a href="https:&#x2F;&#x2F;datadreamer.dev&#x2F;docs&#x2F;latest&#x2F;" rel="nofollow">https:&#x2F;&#x2F;datadreamer.dev&#x2F;docs&#x2F;latest&#x2F;</a><p>Here&#x27;s an (now slightly outdated) example of OpenAI GPT-4 =&gt; OpenAI GPT-3.5: <a href="https:&#x2F;&#x2F;datadreamer.dev&#x2F;docs&#x2F;latest&#x2F;pages&#x2F;get_started&#x2F;quick_tour&#x2F;openai_distillation.html" rel="nofollow">https:&#x2F;&#x2F;datadreamer.dev&#x2F;docs&#x2F;latest&#x2F;pages&#x2F;get_started&#x2F;quick_...</a><p>But you can also do GPT-4 to any model on HuggingFace. Or something like Llama-70B to Llama-1B.<p>For some tasks, this kind of distillation works extremely well given even a few hundred examples of the larger model performing the task.
评论 #42010066 未加载
Permik7 months ago
Do note that this article was posted October 1, 2024, so this capability has been available for a month.
serjester7 months ago
To anyone that thinks models are going to be commodified, it seems like it&#x27;s going to be exceedingly difficult to compete with OpenAI. The developer experience working with them is just too good.<p>Sure you could use a different providers, but you&#x27;re going to be stuck with an incredibly fragmented ops stack. My experience with Google has been shockingly bad and Anthropic has got a good amount of catching up to do. No one else is remotely competitive. Honestly would love to see something from Meta long term.
janalsncm7 months ago
Part of the point of distilling from their models is that I control the model, its availability, and its cost to me. So while this may be a convenient feature, unless I can download the weights it wouldn’t replace my workflow.<p>This does raise the bar for any future startups though. If your plan was to distill GPT4 outputs and lease the weights to me through a REST API, I probably won’t be interested.