cs.CL, cs.LG

Measuring and Mitigating Toxicity in Large Language Models: A Comprehensive Replication Study

arXiv:2605.14087v1 Announce Type: cross
Abstract: Large Language Models (LLMs), when trained on web-scale corpora, inherently absorb toxic patterns from their training data. This leads to “toxic degeneration” where even innocuous prompts can trigger…