KEY LEARNINGS
  • Large Language Models predict likely next words based on patterns, not genuine understanding.
  • Hallucinations are inherent to how LLMs work, not bugs that can be fully eliminated.
  • Sycophancy causes LLMs to prioritize user agreement over accuracy.
  • Retrieval-Augmented Generation can reduce hallucinations by grounding responses in verified sources.
  • Risk-based governance should match controls to the stakes of each deployment use case.
  • Vaswani, A., et al. (2017). Attention Is All You Need. NeurIPS.
  • Brown, T., et al. (2020). Language Models are Few-Shot Learners. NeurIPS.
  • Bai, Y., et al. (2022). Constitutional AI: Harmlessness from AI Feedback. Anthropic.