TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

GLM-4-9B: open-source model with superior performance to Llama-3-8B

66 点作者 marcelsalathe12 个月前

8 条评论

ilaksh12 个月前
Looks like terrific technology. However, the translation says that it's an "irrevocable revocable" non-commercial license with a form to apply for commercial use.
评论 #40597125 未加载
评论 #40606248 未加载
great_psy12 个月前
I’m excited to hear work is being done on models that support function calling natively.<p>Does anybody know if performance could be greatly increased if only a single language was supported ?<p>I suspect there’s a high demand for models that are maybe smaller and can run faster if the tradeoff is support for only English.<p>Is this available in ollama ?
评论 #40594780 未加载
abrichr12 个月前
&gt; GLM-4V-9B possesses dialogue capabilities in both Chinese and English at a high resolution of 1120*1120. In various multimodal evaluations, including comprehensive abilities in Chinese and English, perception &amp; reasoning, text recognition, and chart understanding, GLM-4V-9B demonstrates superior performance compared to GPT-4-turbo-2024-04-09, Gemini 1.0 Pro, Qwen-VL-Max, and Claude 3 Opus.<p>But according to their own evaluation further down, gpt-4o-2024-05-13 outperforms GLM-4V-9B on every task except OCRBench.
评论 #40602750 未加载
norwalkbear12 个月前
Isnt 3-70b so good, reddit llamaers are saying people should buy hardware to run it?<p>Llama-3-8b was garbage for me but damn 70b is good enough
评论 #40603410 未加载
oarth12 个月前
If those numbers are true then it&#x27;s very impressive. Hoping for llama.cpp support.
nubinetwork12 个月前
1M context, but does it really? I&#x27;ve been hit with 32K models that crap out after 10K before...
fragmede12 个月前
model available, not open source.
refulgentis12 个月前
Ehhh man this is frustrating, 7B was a real sweet spot for hobbyist. 8B...doable. I&#x27;ve been joking to myself&#x2F;simultaneously worried that Llama 3 8B and Phi-3 &quot;3B&quot; (3.8B) would start a &quot;ehhh, +1, might as well be a rounding error&quot; thing. It&#x27;s a big deal! I measure a 33% decrease just going from 3B to 3.8B when inferencing on CPU.