Why Does Natural Language Create Hallucinations?
Hallucination is not an LLM bug — it is a structural inevitability of natural language. Four flaws — ambiguity, absent sources, absent confidence, absent time — make it unfixable by scaling alone.
GEUL — A semantically-aligned artificial language for unambiguous communication between humans and AI
Hallucination is not an LLM bug — it is a structural inevitability of natural language. Four flaws — ambiguity, absent sources, absent confidence, absent time — make it unfixable by scaling alone.
Programming languages describe procedures. They cannot describe the world. JSON provides structure but no meaning. Even LISP merely borrows syntax.
Rearranging embedding vectors breaks the model. Avoiding breakage means rebuilding the model from scratch. What we need is not transparency inside the black box, but a transparent layer outside it.
Artificial languages for humans failed — artificial languages for AI are different
GEUL does not reject Wikidata. It transforms the classification system and frequency statistics of 100 million entities into SIDX codebooks. Grammar is built on top of a dictionary.