TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Decoding LLM Uncertainties for Better Predictability

16 点作者 shayanjm超过 1 年前

2 条评论

armcat超过 1 年前
Great work! I love the use of simple techniques like normalized entropy and cosine distance to reveal what the model is "thinking". The example with random number generation is very cool! I actually managed to get that example to work by telling it that it's allowed to use replacement, AND by giving it an example of repeated numbers (just telling it that it can repeat the numbers won't work). Then I get perfect uniform distribution on each step (spikes are all same length). I can definitely see how something like this could be used to guide prompt engineering strategies.
shayanjm超过 1 年前
Building off our last research post, we wanted to figure out ways to quantify &quot;ambiguity&quot; and &quot;uncertainty&quot; in prompts&#x2F;responses to LLMs. We ended up discovering two useful forms of uncertainty: &quot;Structural&quot; and &quot;Conceptual&quot; uncertainty.<p>In a nutshell: Conceptual uncertainty is when the model isn&#x27;t sure what to say, and Structural uncertainty is when the model isn&#x27;t sure how to say it.<p>You can play around with this yourself in the demo!