cs.AI, cs.MA

GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs

arXiv:2604.23366v1 Announce Type: new
Abstract: Autonomous multi-agent LLM systems are increasingly deployed to investigate operational incidents and produce structured diagnostic reports. Their trustworthiness hinges on whether each claim is grounded…