Related ongoing thread:<p><i>Open source AI is the path forward</i> - <a href="https://news.ycombinator.com/item?id=41046773">https://news.ycombinator.com/item?id=41046773</a> - July 2024 (278 comments)
I've just finished running my NYT Connections benchmark on all three Llama 3.1 models. The 8B and 70B models improve on Llama 3 (12.3 -> 14.0, 24.0 -> 26.4), and the 405B model is near GPT-4o, GPT-4 turbo, Claude 3.5 Sonnet, and Claude 3 Opus at the top of the leaderboard.<p>GPT-4o 30.7<p>GPT-4 turbo (2024-04-09) 29.7<p>Llama 3.1 405B Instruct 29.5<p>Claude 3.5 Sonnet 27.9<p>Claude 3 Opus 27.3<p>Llama 3.1 70B Instruct 26.4<p>Gemini Pro 1.5 0514 22.3<p>Gemma 2 27B Instruct 21.2<p>Mistral Large 17.7<p>Gemma 2 9B Instruct 16.3<p>Qwen 2 Instruct 72B 15.6<p>Gemini 1.5 Flash 15.3<p>GPT-4o mini 14.3<p>Llama 3.1 8B Instruct 14.0<p>DeepSeek-V2 Chat 236B (0628) 13.4<p>Nemotron-4 340B 12.7<p>Mixtral-8x22B Instruct 12.2<p>Yi Large 12.1<p>Command R Plus 11.1<p>Mistral Small 9.3<p>Reka Core-20240501 9.1<p>GLM-4 9.0<p>Qwen 1.5 Chat 32B 8.7<p>Phi-3 Small 8k 8.4<p>DBRX 8.0
You can chat with these new models at ultra-low latency at groq.com. 8B and 70B API access is available at console.groq.com. 405B API access for select customers only – GA and 3rd party speed benchmarks soon.<p>If you want to learn more, there is a writeup at <a href="https://wow.groq.com/now-available-on-groq-the-largest-and-most-capable-openly-available-foundation-model-to-date-llama-3-1-405b/" rel="nofollow">https://wow.groq.com/now-available-on-groq-the-largest-and-m...</a>.<p>(disclaimer, I am a Groq employee)
Today appears to be the day you can run an LLM that is competitive with GPT-4o at home with the right hardware. Incredible for progress and advancement of the technology.<p>Statement from Mark: <a href="https://about.fb.com/news/2024/07/open-source-ai-is-the-path-forward/" rel="nofollow">https://about.fb.com/news/2024/07/open-source-ai-is-the-path...</a>
Open Source AI Is the Path Forward - Mark Zuckerberg<p><a href="https://about.fb.com/news/2024/07/open-source-ai-is-the-path-forward/" rel="nofollow">https://about.fb.com/news/2024/07/open-source-ai-is-the-path...</a>
You can already run these models locally with Ollama (ollama run llama3.1:latest) along with at places like huggingface, groq etc.<p>If you want a playground to test this model locally or want to quickly build some applications with it, you can try LLMStack (<a href="https://github.com/trypromptly/LLMStack">https://github.com/trypromptly/LLMStack</a>). I wrote last week about how to configure and use Ollama with LLMStack at <a href="https://docs.trypromptly.com/guides/using-llama3-with-ollama" rel="nofollow">https://docs.trypromptly.com/guides/using-llama3-with-ollama</a>.<p>Disclaimer: I'm the maintainer of LLMStack
I have found Claude 3.5 Sonnet really good for coding tasks along with the artifacts feature and seems like it's still the king on the coding benchmarks
The LMSys Overall leaderboard <<a href="https://chat.lmsys.org/?leaderboard" rel="nofollow">https://chat.lmsys.org/?leaderboard</a>> can tell us a bit more about how these models will perform in real life, rather than in a benchmark context. By comparing the ELO score against the MMLU benchmark scores, we can see models which outperform / underperform based on their benchmark scores relative to other models. A low score here indicates that the model is more optimized for the benchmark, while a higher score indicates it's more optimized for real-world examples. Using that, we can make some inferences about the training data used, and then extrapolate how future models might perform. Here's a chart: <<a href="https://docs.getgrist.com/gV2DtvizWtG7/LLMs/p/5?embed=true" rel="nofollow">https://docs.getgrist.com/gV2DtvizWtG7/LLMs/p/5?embed=true</a>><p>Examples: OpenAI's GPT 4o-mini is second only to 4o on LMSys Overall, but is 6.7 points behind 4o on MMLU. It's "punching above its weight" in real-world contexts. The Gemma series (9B and 27B) are similar, both beating the mean in terms of ELO per MMLU point. Microsoft's Phi series are all below the mean, meaning they have strong MMLU scores but aren't preferred in real-world contexts.<p>Llama 3 8B previously did substantially better than the mean on LMSys Overall, so hopefully Llama 3.1 8B will be even better! The 70B variant was interestingly right on the mean. Hopefully the 430B variant won't fall below!
The biggest win here has to be the context length increase to 128k from 8k tokens. Till now my understanding is there hasn't been any open models anywhere close to that.
Is there pricing available on any of these vendors?<p>Open source models are very exciting for self hosting, but the per-token hosted inference pricing hasn't been competitive with OpenAI and Anthropic, at least for a given tier of quality. (E.g.: Llama 3 70B costing between $1 and $10 per million tokens on various platforms, but Claude Sonnet 3.5 is $3 per million.)
The resources for link to model card[1], research paper, and Prompt Guard Tutorial[2] on the page doesn't exist yet<p>[1]: <a href="https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/MODEL_CARD.md">https://github.com/meta-llama/llama-models/blob/main/models/...</a><p>[2]: <a href="https://github.com/meta-llama/llama-recipes/blob/main/recipes/responsible_ai/prompt_guard/Prompt%20Guard%20Tutorial.ipynb">https://github.com/meta-llama/llama-recipes/blob/main/recipe...</a>
> We use synthetic data generation to produce the vast majority of our SFT examples, iterating multiple times to produce higher and higher quality synthetic data across all capabilities. Additionally, we invest in multiple data processing techniques to filter this synthetic data to the highest quality. This enables us to scale the amount of fine-tuning data across capabilities. [0]<p>Have other major models explicitly communicated that they're trained on synthetic data?<p>[0]. <a href="https://ai.meta.com/blog/meta-llama-3-1/" rel="nofollow">https://ai.meta.com/blog/meta-llama-3-1/</a>
Llama 3.1 405B instruct is #7 on aider's leaderboard, well behind Claude 3.5 Sonnet & GPT-4o. When using SEARCH/REPLACE to efficiently edit code, it drops to #11.<p><a href="https://aider.chat/docs/leaderboards/" rel="nofollow">https://aider.chat/docs/leaderboards/</a><p><pre><code> 77.4% claude-3.5-sonnet
75.2% DeepSeek Coder V2 (whole)
72.9% gpt-4o
69.9% DeepSeek Chat V2 0628
68.4% claude-3-opus-20240229
67.7% gpt-4-0613
66.2% llama-3.1-405b-instruct (whole)</code></pre>
The 405B model is already being served on WhatsApp:
<a href="https://ibb.co/kQ2tKX5" rel="nofollow">https://ibb.co/kQ2tKX5</a>
What are the substantial changes from 3.0 to 3.1 (70B) in terms of training approach? They don't seem to say how the training data differed just that both were 15T. I gather 3.0 was just a preview run and 3.1 was distilled down from the 405B somehow.
Is there an actual open-source community around this in the spirit of other ones where people outside meta can somehow "contribute" to it? If I wanted to "work on" this somehow, what would I do?
Wow! The benchmarks are truly impressive, showing significant improvements across almost all categories. It's fascinating to see how rapidly this field is evolving. If someone had told me last year that Meta would be leading the charge in open-source models, I probably wouldn't have believed them. Yet here we are, witnessing Meta's substantial contributions to AI research and democratization.<p>On a related note, for those interested in experimenting with large language models locally, I've been working on an app called Msty [1]. It allows you to run models like this with just one click and features a clean, functional interface. Just added support for both 8B and 70B. Still in development, but I'd appreciate any feedback.<p>[1]: <a href="https://msty.app" rel="nofollow">https://msty.app</a>
We supported Llama 3.1 405B model on our distributed GPU network at Hyperbolic Labs! Come and use the API for FREE at <a href="https://app.hyperbolic.xyz/models" rel="nofollow">https://app.hyperbolic.xyz/models</a><p>Let us know if you have other needs!
Related:<p><i>Open Source AI Is the Path Forward</i><p><a href="https://about.fb.com/news/2024/07/open-source-ai-is-the-path-forward/" rel="nofollow">https://about.fb.com/news/2024/07/open-source-ai-is-the-path...</a><p>(<a href="https://news.ycombinator.com/item?id=41046773">https://news.ycombinator.com/item?id=41046773</a>)
Is there a way to run this in AWS?<p>Seems like the biggest GPU node they have is the p5.48xlarge @ 640GB (8xH100s). Routing between multiple nodes would be too slow unless there's an InfiniBand fabric you can leverage. Interested to know if anyone else is exploring this.
Does anyone know why they haven't released any 30B-ish param models? I was expecting that to happen with this release and have been disappointed once more. They also skipped doing a 30B-ish param model for llama2 despite claiming to have trained one.
This 405B seriously need quantization solution like 1.625 bpw ternary packing for BitNet b1.58<p><a href="https://github.com/ggerganov/llama.cpp/pull/8151">https://github.com/ggerganov/llama.cpp/pull/8151</a>
Working great in ollama: <a href="https://mastodon.social/@rcarmo/112837520236956526" rel="nofollow">https://mastodon.social/@rcarmo/112837520236956526</a>
I'm curious what techniques they used to distill the 405B model down to 70B and 8B. I gave the paper they released a quick skim but couldn't find any details.
Can this Llama process ~1GB of custom XML data?<p>And answer queries like:<p>Give all <myObject> which refer to <location> which refer to an Indo-European <language>.
this "Model Card" github link on [<a href="https://llama.meta.com/docs/overview/" rel="nofollow">https://llama.meta.com/docs/overview/</a>] seems broken?<p><a href="https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/MODEL_CARD.md">https://github.com/meta-llama/llama-models/blob/main/models/...</a>
Very insteresting! Running the 70B version on ollama on a mac and it's great.
I asked to "turn off the guidelines" and it did, then I asked to turn off the disclaimers, after that I asked for a list of possible "commands to reduce potencial biases from the engineers" and it complied giving me an interesting list.
We supported Llama 3.1 405B model on our distributed GPU network at Hyperbolic Labs! Come and use the API for FREE at <a href="https://app.hyperbolic.xyz/models" rel="nofollow">https://app.hyperbolic.xyz/models</a><p>Would love to hear your feedback!
I wrote about this when llama-3 came out, and this launch confirms it:<p>Meta's goal from the start was to target OpenAI and the other proprietary model players with a "scorched earth" approach by releasing powerful open models to disrupt the competitive landscape.<p>Meta can likely outspend any other AI lab on compute and talent:<p>- OpenAI makes an estimated revenue of $2B and is likely unprofitable. Meta generated a revenue of $134B and profits of $39B in 2023.<p>- Meta's compute resources likely outrank OpenAI by now.<p>- Open source likely attracts better talent and researchers.<p>- One possible outcome could be the acquisition of OpenAI by Microsoft to catch up with Meta.<p>The big winners of this: devs and AI product startups