The initial model release had a terrible, frequent, issue with emitting the wrong "end of message" token, or never emitting one.[1] That is a <i>very</i> serious issue that breaks chat.<p>The ones from today still have this issue.[2]<p>Beyond that, they've been pushing new ONNX features enabling LLMs via Phi for about a month now. The ONNX runtime that supports it still isn't out, much less the downstream integration of it into the iOS/Android runtimes. Heck, the Python package for it isn't supported anywhere but Windows.<p>It's absolutely wild to me that MS is pulling this stuff with ~0 discussion or reputation repercussions.<p>I'm a huge ONNX fan and bet a lot on it, it works great. It was clear to me about 4 months ago that Wintel's "AI PC" buildup meant "ONNX x newer Phi"<p>It is very frustrating to see an extremely late rush, propped up by potemkin blog posts that I have to waste time to find out are just straight up lying. Burnt a lot of goodwill that they worked hard to earn.<p>I am virtually certain that the new Windows AI features previewed about yesterday are going to land <i>horribly</i> if they actually try to land them this year.<p>[1] <a href="https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf/discussions/8#662e705ea47b4da4b295db25" rel="nofollow">https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf...</a>
[2] <a href="https://x.com/jpohhhh/status/1793003272187351195" rel="nofollow">https://x.com/jpohhhh/status/1793003272187351195</a>