TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Qualcomm works with Meta to enable on-device AI applications using Llama 2

104 pointsby ahiknsralmost 2 years ago

9 comments

behnamohalmost 2 years ago
Meanwhile Apple is stubbornly insisting on its own ways, and awkwardly silent during this whole AI revolution. I gave up using Siri years ago due to its glaring stupidity compared with Google Assistant and Alexa.<p>While Apple keeps making money from overpriced hardware, the competitors are working on actually being pioneers in AI. It makes me sad to see so much computational power in my iPhone and iPad getting wasted on silly subpar iOS apps.
评论 #36779155 未加载
评论 #36779169 未加载
评论 #36780302 未加载
评论 #36779452 未加载
评论 #36779117 未加载
评论 #36780235 未加载
评论 #36779140 未加载
评论 #36779621 未加载
评论 #36781184 未加载
评论 #36779120 未加载
评论 #36779181 未加载
simonwalmost 2 years ago
If you have a modern iPhone you can try running an LLM directly on it today using the MLC iPhone app: <a href="https:&#x2F;&#x2F;mlc.ai&#x2F;mlc-llm&#x2F;#iphone" rel="nofollow noreferrer">https:&#x2F;&#x2F;mlc.ai&#x2F;mlc-llm&#x2F;#iphone</a><p>It can run Vicuna-7B which is a pretty impressive model.<p>(They have an Android app too but I haven&#x27;t tried that yet).
评论 #36779395 未加载
评论 #36781194 未加载
smoldesualmost 2 years ago
&gt; The ability to run generative AI models like Llama 2 on devices such as smartphones, PCs, <i>VR&#x2F;AR headsets</i><p>Maybe it&#x27;s an upcoming feature for the Quest 3?<p>To that end, I&#x27;ve been pretty amazed by how far quantization has come. Some early llama-2 quantizations[0] have gotten down to ~2.8gb, though I haven&#x27;t tested it to see how it performs yet. Still though, we&#x27;re now talking about models that can comfortably run on pretty low-end hardware. It will be interesting to see where llama crops up with so many options for inferencing hardware.<p>[0] <a href="https:&#x2F;&#x2F;huggingface.co&#x2F;TheBloke&#x2F;Llama-2-7B-Chat-GGML&#x2F;tree&#x2F;main" rel="nofollow noreferrer">https:&#x2F;&#x2F;huggingface.co&#x2F;TheBloke&#x2F;Llama-2-7B-Chat-GGML&#x2F;tree&#x2F;ma...</a>
m3kw9almost 2 years ago
ChatGPT 3.5 is the base level people expect LLMs to be, it would be 2-3 generation(3-4 years) of hardware before we can reach that. Anything below is just going to get bad reviews
评论 #36778935 未加载
评论 #36779133 未加载
评论 #36779005 未加载
评论 #36781199 未加载
评论 #36779163 未加载
transcriptasealmost 2 years ago
Facebook hardware in my devices that claims it’s there to protect my privacy.<p>What’s the catch? Firmware based “anonymous” telemetry?
评论 #36780290 未加载
评论 #36779515 未加载
seydoralmost 2 years ago
FB should release a phone, they already have quest OS
评论 #36779268 未加载
评论 #36779354 未加载
ssss11almost 2 years ago
Hmm I wonder what sort of privacy impacts there are in a future of having Meta (or Google et al) AI running on chips on your phone when the parent company has so much info on you and blatantly flaunts privacy laws.
vorpalhexalmost 2 years ago
How would it work to get a model that needs 8Gb+ of vram currently into some chiplet form factor? Is there an obvious way of translating this more directly to hardware?
评论 #36779903 未加载
评论 #36779205 未加载
评论 #36779033 未加载
oneplanealmost 2 years ago
I doubt Qualcomm will be able to increase their performance ahead of Nvidia decreasing their energy requirements.
评论 #36779097 未加载
评论 #36779867 未加载