Extrinsic Hallucinations in LLMs

Date: July 7, 2024 | Estimated Reading Time: 29 min | Author: Lilian Weng

Hallucination in large language models usually refers to the model generating unfaithful, fabricated, inconsistent, or nonsensical content. As a term, hallucination has been somewhat generalized to case...