TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: How are you making sure LLM provider doesn't ban you from user prompts?

1 pointsby spxneoabout 1 year ago
One of my worry is my users submitting malicious prompts and getting my account banned for everybody.<p>How do you ensure that the prompt doesn&#x27;t violate LLM providers Terms of Service?<p>Does this mean we have to call the text gen API twice? First to analyze and ask if it violates TOS and then second to finally fulfill the user&#x27;s requests?<p>Are there any other solutions that is working for you?

3 comments

dragonwriterabout 1 year ago
&gt; Does this mean we have to call the text gen API twice? First to analyze and ask if it violates TOS and then second to finally fulfill the user&#x27;s requests?<p>This just means the user generated content gets sent to the API once with different framing (risking a ban or strike or whatever) and if it doesn&#x27;t trigger your detection gets sent again with the normal framing (giving another chance at a provider ban, strike, etc.)<p>Seems like that would just accelerate your ban by having you send each potentially-violating interaction twice, with slightly different context, giving more chances of a violation and possibly doubling violations for some content.<p>You can probably do better at reducing your risk by running a local classifier (or a comparatively small local LLM) as your trouble detector, before deciding to send a request to the backend, though validating the trouble detector setup may be problematic.
评论 #40293726 未加载
mepianabout 1 year ago
You can use a local LLM like Llama 3, they&#x27;re pretty good. Zero risk of getting banned by them.
JojoFatsaniabout 1 year ago
Wouldn’t the “check” api call be enough to violate TOC?
评论 #40293729 未加载