GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs
arXiv:2604.23366v1 Announce Type: new
Abstract: Autonomous multi-agent LLM systems are increasingly deployed to investigate operational incidents and produce structured diagnostic reports. Their trustworthiness hinges on whether each claim is grounded…