The most frustrating thing about the many, <i>many</i> clones of this exact type of idea is that pretty much <i>all</i> of them require OpenAI.<p>Stop doing that.<p>You will have way more users if you make OpenAI (or anything that requires cloud) the 'technically possible but pretty difficult art of hoops to make it happen' option, instead of the other way around.<p>The best way to make these apps IMO is to make them work <i>entirely</i> locally, with an easy string that's swappable in a .toml file to any huggingface model. Then if you <i>really</i> want OpenAI crap, you can make it happen with some other docker secret or `pass` chain or something with a key, while changing up the config.<p>The default should be local first, do as much as possible, and then <i>if the user /really/ wants to</i>, make the collated prompt send a very few set of tokens to openAI.
Keep your data private and don't leak it to third parties. Use something like privateGPT (32k stars). Not your keys, not your data.<p>"Interact privately with your documents using the power of GPT, 100% privately, no data leaks"[0]<p>[0] <a href="https://github.com/imartinez/privateGPT">https://github.com/imartinez/privateGPT</a>
Is it going to send my personal data to OpenAI? Isn't that a serious problem? Does not sound like a wise thing to do, not at least without redacting all sensitive personal data from the data. Am I missing something?
This readme is very confusing. It says we're going to use the GPT-2 tokenizer, and use GPT-2 as an embedding model. But looking at the code, it seems to use the default LangChain OpenAIEmbeddings and OpenAI LLM. Aren't those text-embedding-ada-002 and text-davinci-003, respectively?<p>I don't understand how GPT-2 enters into this at all.
I don't get it, GPT-2 is (one of the few) open models from OpenAI, you can just run it locally, why would you use their API for this?
<a href="https://github.com/openai/gpt-2">https://github.com/openai/gpt-2</a>
Anyone know how milvus, quickwit, pinecone compares?<p>I've been thinking about seeing if there's consulting opportunities for local businesses for LLMs, finetuning/vector search, chat bots. Also making tools to make it easier to drag and drop files and get personalized inference. Recently I saw this one pop into my linkedin feed, <a href="https://gpt-trainer.com/" rel="nofollow noreferrer">https://gpt-trainer.com/</a> . There's been a few others for documents I've found<p><a href="https://www.explainpaper.com/" rel="nofollow noreferrer">https://www.explainpaper.com/</a><p><a href="https://www.konjer.xyz/" rel="nofollow noreferrer">https://www.konjer.xyz/</a><p>Nope nope, wouldn't want to compete with that on pricing. Local open source LLMs on a 3090 would also be a cool service, but wouldn't have any scalability.<p>Are there any other finetuning or vector search context startups you've seen?
I’m working for a company that works as a security layer between any sensitive enterprise data and the LLMs. Regardless of the model (HF, ChatGPT, Bard), and regardless of the medium - conversational data, pdf, knowledge bases like Notion etc. It hides the sensitive data, preventing risky use and fact checking at the same time. Happy to make an intro if that’s what you’re looking for! tothepoint.tech
Don't build a personal ChatGPT, and don't let OpenAI, Microsoft and their business partners (and probably the US government) have a bunch of your personal and private information.
Please provide this reference in your readme / blog as it is the original source for your work... and provides the background for the tradeoff between the 2 approaches: 1) fine-tuning vs 2) Search-ask<p><a href="https://github.com/openai/openai-cookbook/blob/main/examples/Question_answering_using_embeddings.ipynb">https://github.com/openai/openai-cookbook/blob/main/examples...</a>
The author has a demo of this here: <a href="https://www.swamisivananda.ai/" rel="nofollow noreferrer">https://www.swamisivananda.ai/</a>