cs.CL

Generating Effective CoT Traces for Mitigating Causal Hallucination

arXiv:2604.12748v1 Announce Type: new
Abstract: Although large language models (LLMs) excel in complex reasoning tasks, they suffer from severe causal hallucination in event causality identification (ECI), particularly in smaller models ($\leq$1.5B pa…