Extrinsic Hallucinations in LLMs
Date: July 7, 2024 | Estimated Reading Time: 29 min | Author: Lilian Weng
Hallucination in large language models usually refers to the model generating unfaithful, fabricated, inconsistent, or nonsensical content. As a term, hallucination has been somewhat generalized to case...