Extrinsic Hallucinations in LLMs
Hallucination in large language models usually refers to the model generating unfaithful, fabricated, inconsistent, or nonsensical content. As a term, hallucination has been somewhat generalized to case...
Date: July 7, 2024|Estimated Reading Time: 29 min|Author: Lilian Weng