I’m curious: does this fundamentally <i>need</i> to contain an actual model, or would it be okay if it generated a synthetic model itself, full of random weights? I’m picturing downloading just, say, a 20MB file instead of the multi-gigabyte one, and…<p>Hang on, why is <a href="https://blob.localscore.ai/localscore-0.9.2" rel="nofollow">https://blob.localscore.ai/localscore-0.9.2</a> 380MB? I remember llamafile being only a few megabytes. From <a href="https://github.com/Mozilla-Ocho/llamafile/releases">https://github.com/Mozilla-Ocho/llamafile/releases</a>, looks like it steadily grew from adding support for GPUs on more platforms, up to 28.5MiB¹ in 0.8.12, and then rocketed up to 230MiB in 0.8.13:<p>> <i>The llamafile executable size is increased from 30mb to 200mb by this release. This is caused by <a href="https://github.com/ggml-org/llama.cpp/issues/7156">https://github.com/ggml-org/llama.cpp/issues/7156</a>. We're already employing some workarounds to minimize the impact of upstream development contributions on binary size, and we're aiming to find more in the near future.</i><p>Ah, of course, <i>CUDA</i>. Honestly I might be more surprised that it’s <i>only</i> this big. That monstrosity will happily consume a dozen gigabytes of disk space.<p>llamafile-0.9.0 was still 231MiB, then llamafile-0.9.1 was 391MiB, now llamafile-0.9.2 is 293MiB. Fluctuating all over the place, but growing a lot. And localscore-0.9.2 is 363MiB. Why 70MiB extra on top of llamafile-0.9.2? I’m curious, but not curious enough to investigate concretely.<p>Well, this became a grumble about bloat, but I’d still like to know whether it would be feasible to ship a smaller localscore that would synthesise a suitable model, according to the size required, at runtime.<p>—⁂—<p>¹ Eww, GitHub is using the “MB” suffix for its file sizes, but they’re actually mebibytes (2²⁰ bytes, 1048576 bytes, MiB). I thought we’d basically settled on returning the M/mega- prefix to SI with its traditional 10⁶ definition, at least for file sizes, ten or fifteen years ago.