TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: Chat with multiple LLMs: o1-high-effort, Sonnet 3.5, GPT-4o, and more

62 pointsby cr4zy4 months ago
Hello HN! I was fed up switching between multiple UIs to ask GPT, Claude, etc… the same question and comparing the answers. So I built a way to ask multiple models the same question efficiently by having the LLM compare the responses and only show you new and valuable information from the 2nd model. This way you still get a fast response as normal from the 1st model, but also get any added value provided by the 2nd model.<p>Initially I built my own UI to use this, but stumbled upon Open WebUI (formerly Ollama WebUI) which is fantastic, but is made more for local access to LLMs. So I talked to its creator, Timothy Baek, and he mentioned that security needed to be shored up before production deployment. I did some scans with semgrep, and fixed some XSRF and CORS issues along with making sure the JWT tokens, passwords, etc… were secure. This was in addition to other folks&#x27; amazing security contributions. So now we build on Open WebUI! oss ftw<p>After launching privately a month or so ago and posting a few reddit links, I have about 100 users.<p>We also just got access to the o1 API and provide o1-high, medium, and low effort. o1 high effort is able to solve coding problems that only o1 pro can also solve, though pro can go further and has better formatting, o1 high effort is a nice option if you don&#x27;t want to fork over $200&#x2F;mo.<p><a href="https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;OpenAI&#x2F;comments&#x2F;1i0r867&#x2F;comment&#x2F;m70xlsk&#x2F;?context=3" rel="nofollow">https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;OpenAI&#x2F;comments&#x2F;1i0r867&#x2F;comment&#x2F;m70...</a><p>You can use o1 for free in PolyChat, which is the only place I&#x27;ve seen you can do so.<p>You can also ask multiple models the same question and stream the answers simultaneously side by side. <a href="https:&#x2F;&#x2F;imgur.com&#x2F;a&#x2F;ly5ZPh8" rel="nofollow">https:&#x2F;&#x2F;imgur.com&#x2F;a&#x2F;ly5ZPh8</a><p>And you can have multiple chats going simultaneously and they will continue in the background and notify you when they&#x27;re done.<p>Another cool feature that makes long chats way easier to navigate is the overview (three dots on top right of chat) <a href="https:&#x2F;&#x2F;imgur.com&#x2F;a&#x2F;PoB1phT" rel="nofollow">https:&#x2F;&#x2F;imgur.com&#x2F;a&#x2F;PoB1phT</a><p>We give you full control, unlike most providers, to change the system prompt, temperature, etc.. in the chat settings in the &quot;Controls&quot; on the top right.<p>The formatting for code outputs and inputs is fantastic, using codemirror, and you can run code in the code blocks if it&#x27;s python or JS using in-browser runtimes.<p>You can also share chats within polychat where other logged in users can see them:<p>e.g. <a href="https:&#x2F;&#x2F;polychat.co&#x2F;s&#x2F;ec163812-9323-4cb5-a429-7ddeaf3572c3" rel="nofollow">https:&#x2F;&#x2F;polychat.co&#x2F;s&#x2F;ec163812-9323-4cb5-a429-7ddeaf3572c3</a><p>Or you can share it publicly to the OpenWeb UI community<p><a href="https:&#x2F;&#x2F;openwebui.com&#x2F;c&#x2F;crizcraig&#x2F;e1b2fff9-30a3-4ee7-8eed-b243bbc73b41" rel="nofollow">https:&#x2F;&#x2F;openwebui.com&#x2F;c&#x2F;crizcraig&#x2F;e1b2fff9-30a3-4ee7-8eed-b2...</a><p>Finally we allow you to search and organize your chats into folders which makes finding things super fast!<p>Infra: My custom backend that combines models and does things Open WebUI doesn&#x27;t handle like token tracking is written in FastAPI and uses LiteLLM for easily accessing different model APIs. I host everything on GCP using Cloud Run for the backend and use PostGres for Open WebUI&#x27;s db, along with BigQuery and Firestore for my FastAPI app&#x27;s db.<p>Pricing: Our pricing makes it cheap to access top models. It&#x27;s free at first, then we offer subscription tiers starting at $5&#x2F;mo which allots about ~1 million tokens per month, enough for most people. But unlike Claude&#x27;s UI for example, we don&#x27;t shut you down with rate limits, but rather let heavy users upgrade beyond to $10, $20, $40, $80, etc... We also suggest a plan based on how quickly you used your free tokens, so you have an idea upfront of the monthly cost before you buy. And you can upgrade or downgrade at any time.<p>Thanks Show HN &lt;3

15 comments

hoerzu4 months ago
Created an open source alternative just using the browser and not sharing your data with a third party: <a href="https:&#x2F;&#x2F;chromewebstore.google.com&#x2F;detail&#x2F;tabgpt-ask-chatgpt-gemini&#x2F;aedecpiooconcmeacmocfalcpjomield" rel="nofollow">https:&#x2F;&#x2F;chromewebstore.google.com&#x2F;detail&#x2F;tabgpt-ask-chatgpt-...</a>
评论 #42809582 未加载
pogue4 months ago
Fyi: Duckduckgo has a free ai chat with o1-mini at <a href="https:&#x2F;&#x2F;duck.ai" rel="nofollow">https:&#x2F;&#x2F;duck.ai</a><p>They also offer Claude 3 Haiku, Llama 3.1 70B and Mixtral 8x7B. No account required and they say it&#x27;s not logged + anonymous. I&#x27;m sure they have limits but I never reached them.
评论 #42811613 未加载
oidar4 months ago
This is very good at this right now too: <a href="https:&#x2F;&#x2F;melty.sh&#x2F;chorus">https:&#x2F;&#x2F;melty.sh&#x2F;chorus</a>
评论 #42810699 未加载
评论 #42809351 未加载
artdigital4 months ago
Some other options:<p>OpenRouter has similar features with any model that’s available on OpenRouter through open router billing<p>On the desktop, Msty can do this too
victorantos4 months ago
I’ve built a multi chat web app as well, atm offering 3 simultaneous chats: ChatGPT, Claude and Gemini<p>My USP is that you can generate and share a n image of the chats, more for Premium users.<p><a href="https:&#x2F;&#x2F;sneos.com" rel="nofollow">https:&#x2F;&#x2F;sneos.com</a>
nemoniac4 months ago
gptel can do this already with over 20 models.<p>Costs nothing beyond your subscriptions to the individual models.<p><a href="https:&#x2F;&#x2F;github.com&#x2F;karthink&#x2F;gptel">https:&#x2F;&#x2F;github.com&#x2F;karthink&#x2F;gptel</a><p>The author shows it with 4 models at once in this demo video:<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=bsRnh_brggM&amp;t=853s" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=bsRnh_brggM&amp;t=853s</a>
victorantos4 months ago
Looks great! I’m working on something similar: <a href="https:&#x2F;&#x2F;sneos.com" rel="nofollow">https:&#x2F;&#x2F;sneos.com</a>. Planning to launch on HN once I nail down the scaling and usage limits—currently free for everyone. Would love to hear how your journey with Polychat is going!
olddog24 months ago
Cool, but nano-gpt.com do pay per use for about 100 chat and image models including o1pro, deepseek r1 etc.
评论 #42809256 未加载
wslh4 months ago
We have Geppetto, a FOSS that you can ask questions within Slack and using several services. Feel free to fork it: &lt;<a href="https:&#x2F;&#x2F;github.com&#x2F;Deeptechia&#x2F;geppetto">https:&#x2F;&#x2F;github.com&#x2F;Deeptechia&#x2F;geppetto</a>&gt;
cr4zy4 months ago
So it looks like some folks are getting errors with the non-streaming models, i.e. the o1 models. I think their long running cxns with zero packets may cause some networks to drop the requests. Will look into a hearbeat&#x2F;keepalive on those.
评论 #42819191 未加载
liberix4 months ago
I&#x27;ve noticed that you offer an API key under Settings &#x2F; Account. How does the API work? Is there any documentation? I&#x27;d also like to see the pricing details for the different plans.
评论 #42819180 未加载
ilrwbwrkhv4 months ago
If it doesn&#x27;t have DeepSeek R1 it is a non starter.
moralestapia4 months ago
Unlimited free OpenAI o1?<p>OP, isn&#x27;t that really expensive to maintain?
评论 #42810747 未加载
derekhsu4 months ago
Seriously? Run a Open WebUI and get money?
Frummy4 months ago
Man if this idea makes &gt;500 bucks per month no offense I mean it positively you just motivated me to unironically create my somewhat as simple idea that was laughed at by literally everyone in a geek discord Like maybe the lowhanging fruit are actually hanging low
评论 #42809553 未加载
评论 #42809572 未加载
评论 #42818227 未加载