Concept Training for Human-Aligned Language Models
arXiv:2603.29123v1 Announce Type: new
Abstract: The next-token prediction (NTP) objective trains language models to predict a single continuation token at each step. In natural language, however, a prefix can be continued in many valid ways, and even …