Researchers conducted an audit of citation fabrication by large language models across four academic domains, finding hallucination rates vary widely depending on model and prompt framing. Key takeaway for content creators: using multi-model consensus and within-prompt repetition can significantly improve accuracy in detecting fabricated citations.
Read the full article at arXiv cs.CL (NLP)
Want to create content about this topic? Use Nemati AI tools to generate articles, social posts, and more.





