TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Sharing learnings from evaluating Million+ LLM responses

3 点作者 sourabh03agr超过 1 年前
Hey everyone, we have been working with many LLM builders to help them evaluate and improve the quality of their LLM applications and recently penned down our learnings on how exactly to evaluate them. We have defined multiple evaluation criteria and grouped them according to their applicability. We have the full blog on our website (https:&#x2F;&#x2F;uptrain.ai&#x2F;blog&#x2F;how-to-evaluate-your-llm-applications) and am just copying the key excerpts from it. Hope you find it useful.<p>Why is Evaluating LLMs Tricky?<p>LLMs are great but they introduce novel complexities when it comes to measuring their performance. In traditional ML&#x2F;DL, we can annotate data and have “Ground Truths” to compute scores like precision, recall, and accuracy of the model. LLMs on the other hand, operate in realms where it is impossible to define a clear and unique “Ground Truth” to do a word-to-word comparison against. There could be 100s of different email copies and all of them could be equally correct for the task.<p>When traditional NLP metrics fail, novel techniques have emerged to use LLMs as evaluators. But how can one expect for GPT to reliably score its own response - too self-consumed?<p>The trick is to simplify the evaluation task. Instead of directly asking whether the response is correct or not, ask the LLM to evaluate the response only along a certain dimension - is it grounded by the context, is it too verbose, is it answering all aspects of the question, is the tone correct, etc.<p>Now, the dimensions one should care about depend upon the use case. For applications generating marketing copies, tonality, interestingness, etc. would matter more while for RAG and chatbots, factual accuracy, retrieval quality, and response relevance would be better to look at.<p>Dimensions of LLM Evaluations<p>Taking our learnings while building UpTrain - an open-source LLM evaluation tool, we have divided these dimensions into 4 categories:<p>1.Checks for evaluating Task Understanding and Context Awareness: Check to evaluate if your LLM + prompt configuration can comprehend the task at hand and fully utilize the provided context to provide an appropriate response. This category is further divided into 2 sub-categories:<p>a. Response Appropriateness: Intent understanding, Response Completeness, Response Relevancy, Output Structure Integrity, etc. b. Context Awareness and Grounding: Hallucinations, Retrieved-context quality, context utilization, etc.<p>2. Checks for evaluating Language Quality: Dimensions that help to evaluate the quality of the response from a language perspective. We further divide this into two subcategories:<p>a. Task Independent: Dimensions like grammar correctness, fluency, coherence, toxicity, fairness towards all sectors of society, etc. We generally see high scores for these, especially for RLHFed LLMs b. Task Dependent: Tonality match with the given persona, creativity, interestingness, etc. Your prompt can play a big role here<p>3. Evaluating Reasoning Capabilities: Includes dimensions like logical correctness (right conclusions), logical robustness (consistent with minor input changes), logical efficiency (shortest solution path), and common sense understanding (grasping common concepts). One can’t do much beyond prompting techniques like CoT and primarily depends upon the LLM chosen.<p>4. Custom Evaluations: Many applications require customized metrics tailored to their specific needs. You want adherence to custom guidelines, check for certain keywords, etc.<p>You can read the full blog here (https:&#x2F;&#x2F;uptrain.ai&#x2F;blog&#x2F;how-to-evaluate-your-llm-applications). Hope you find it useful. I am one of the developer of UpTrain - it is an open-source package to evaluate LLM applications (https:&#x2F;&#x2F;github.com&#x2F;uptrain-ai&#x2F;uptrain).<p>Would love to get feedback from the HN community.

暂无评论

暂无评论