Side note: I feel like ChatGPT's long-term memory isn't implemented properly. If you check the 'saved memories,' they are just bad.
Nice. I think in the future this could be way better if everything was local and didn't require a API key. As far as I can tell mem0 is a fancy retrieval system. It could probably work pretty well locally with simpler models
I'm a long time Claude user.<p>Instead of long-term memory I'd be happy if it had short-term reliability. I lost count the number of times this week that Claude failed to process prompts because it was down.
I always wonder what the heck people are thinking when they invent some cool AI feature and implement it for one specific LLM since we already have the technology/libraries to make most anything you want to do be able to work with most any LLM. (For you pedantic types, feel free to point out the exceptions).<p>Personally I use LangChain/Python for this, and that way any new AI features I create therefore easily work across ALL LLMs, and my app just lets the end user pick the LLM they want to run on. Every feature I have works on every LLM.