Why AI hallucinate and how to reduce it
LLM hallucinate (invents facts that do not exist) because of copyright and AI benchmarks :
AI are built and trained to avoid copyright problems of the data they absorbed at training time. Their fine tuning push them to avoid repetition therefore "invent" something similar to what they learnt
AI benchmarks sanction AI that spit the answer of"classical" problems they were trained on. They are fine tuned to "invent" similar reasoning but not duplicate results.
Understanding why AI hallucinate is helpful to design prompts that will orient AI to not hallucinate. The key is to provide data on which the AI has to work and to limit the perimeter of its answer.