Furthermore, LLMs are prone to "hallucination"—that is, they write convincingly about concepts and sources that are either incorrect or nonexistent. "You can't necessarily trust an LLM to ...