TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Persimmon-8B

175 pointsby jgershenover 1 year ago

14 comments

AaronFrielover 1 year ago
I hope this is only a slight tangent; since the authors talk about their model serving throughput and I hope I can get a gut-check on my understanding of the state-of-the-art of model serving.<p>The success of ChatGPT and my current work has had me thinking a lot about the &quot;product&quot; applications of large language models. I work at Pulumi on www.pulumi.com&#x2F;ai; it&#x27;s a GPT-3.5 and GPT-4 interface using retrieval augmented generation to generate Pulumi programs, and user experience is top of mind for me.<p>(Fingers crossed this doesn&#x27;t hug our site to death here for the reasons I&#x27;m about to explain.)<p>To be blunt: I have found it surprisingly difficult to find the right tools to host models without dramatically worsening the UX. In theory we should be able to fine-tune a model against our own SDKs and synthetically generated code to improve the model&#x27;s output and to guard against hallucination when retrieval fails. In practice, self-hosted model serving APIs have really poor time-to-first-token or even completely lack streaming behavior. It&#x27;s a non-starter to build a product on something where a user has to sit and watch a spinner for a minute or more. I&#x27;ve been looking at the vLLM project with great interest, but haven&#x27;t found much else.<p>---<p>For folks in MLops, deploying models with streaming APIs:<p>1. Is it mostly accurate that none of the model serving tools created prior to ChatGPT are great for streaming, interactive use cases?<p>2. How are you currently serving these models as an API and what upcoming tools are you exploring?<p>For the authors: How does your inference optimization compare to vLLM, or other tools using techniques such as continuous batching and paged attention?
评论 #37432159 未加载
gardnrover 1 year ago
Two important takeaways on the base model:<p>* scored 18.9 on HumanEval (coding) where Llama2 7B scored 12.2<p>* was trained from the beginning with a 16k context using a modified RoPe where many models are simply fine-tuned using RoPe to gain longer context windows after the base model has been trained at 4k.<p>Can anyone share ideas on how important the 2nd one is? Do LLMs benefit from large context windows using RoPe during pretraining?
评论 #37424968 未加载
评论 #37427189 未加载
评论 #37427131 未加载
thewataccountover 1 year ago
Awesome! I applaud everyone training new models and attempting different techniques!<p>I&#x27;m concerned about the current download&#x27;s availability - its two URLs to some object storage. I find that these go dark rather quickly for many different reasons (accidentally moving it, bandwidth limits, deleting it later, etc).<p>I&#x27;m curious if there&#x27;s a reason it&#x27;s not also hosted on huggingface? I&#x27;m not saying they&#x27;re the best place, but redundancy is good, most models have entries there, they have a very good cdn, and isn&#x27;t as likely to go dark accidentally.
评论 #37430283 未加载
评论 #37424881 未加载
123yawaworht456over 1 year ago
I applaud you guys for not including any nauseating gibberish in this press release or seemingly anywhere else on your website. It&#x27;s like a breath of fresh air comparing to every other AI-related resource I saw recently. Please, keep it up.
评论 #37424102 未加载
imjonseover 1 year ago
Congrats on the release! Two questions.<p>1) In the results table, Llama2 base is being compared to Persimmon base and finetuned, and only the latter performs better. Would a comparison to Llama2-chat be possible&#x2F;fair?<p>2) The Llama-2 numbers for MMLU in that table seem different from those in the HF leaderboard and the Llama-2 webpage presentation. Is it the 1-shot variant that is different or are these measurements not 100% standard and reproducible?
评论 #37425609 未加载
elietoubiover 1 year ago
Really cool! Honestly I wish these releases would come with a demo (like on replicate or hugging face)
deckar01over 1 year ago
The docker container fails installing flash-attn… but honestly a giant API container on top of a custom model generation framework loses all the benefits of Torch’s standard interfaces. It doesn’t really matter how optimized your model runtime is if it’s cemented into a synchronous monolith. The metric that should be optimized is time to first decoded token, because that is how speed is perceived by humans reading the output.
评论 #37426146 未加载
automatististover 1 year ago
&gt; The standard practice for achieving fast inference is to rewrite the entire model inference loop in C++, as in FasterTransformer, and call out to special fused kernels in CUDA. But this means that any changes to the model require painfully reimplementing every feature twice: once in Python &#x2F; PyTorch in the training code and again in C++ in the inference codebase. We found this process too cumbersome and error prone to iterate quickly on the model.<p>I am an AI novice but why can&#x27;t they automated this with AI? I thought the whole point of these tools was to automated tasks that are error prone and require lots of attention to details. Computers are great at that kind of stuff so it&#x27;s surprising they haven&#x27;t applied AI techniques to automate parts of the AI pipeline like converting code from Python to C++.
评论 #37424280 未加载
评论 #37423870 未加载
Havocover 1 year ago
&gt;The model has 70k unused embeddings for multimodal extensions,<p>Could someone briefly explain what this means? multimodal as in picture, but if unused then presumably that part is somehow untrained...so it wouldn&#x27;t know what to do with the picture?
评论 #37426655 未加载
TrueDualityover 1 year ago
Appreciate the release! Since you&#x27;re hosting the downloads directly, I&#x27;d recommend throwing an integrity hash for each of the files alongside the download links so users can verify there wasn&#x27;t any corruption in transfer.
评论 #37424888 未加载
评论 #37424848 未加载
rvzover 1 year ago
Good. Keep it going. Let’s have more $0 for free AI models getting released since we all know it is the future and you can’t compete with free.<p>The AI race to zero must be accelerated with $0 free models and less control from gatekeepers such as ClosedAI
sunshadowover 1 year ago
Do you have any explanations on why this performed better than Llama 2?
评论 #37435506 未加载
visargaover 1 year ago
Since it is coming from Adept, maybe they are building 8B models for UI automation, the inputs are usually large and latency required is low. It&#x27;s basically a task of information extraction and UI action generation.
theLiminatorover 1 year ago
What kind of use cases do these sub 10B param models serve? Are they mostly useful for code completion?
评论 #37425724 未加载
评论 #37427329 未加载