It's interesting to see all this hard work being done specifically for "fact-fixing" <i>inside</i> neural networks, whereas I think the future is probably having two models: one for language processing (grammar, etc.) and the other for semantic mapping (where we encode <i>actual</i> relations and properties, causality, etc.). To wit, unless you squint really <i>really</i> hard, this is not exactly true:<p>> Language models can be viewed as knowledge bases containing memorized tuples (s, r, o), each connecting some subject s to an object o via a relation...<p>LLMs don't have the concept of objects or relationships. You might be able to argue some of that ends up being encoded in the embeddings (especially if they're particularly big), but I would posit that those embeddings mostly end up handling the grammar. So "ball" is associated with "red" purely because of locality, but training an <i>actual</i> knowledge base would be much more powerful.