MachineLearning

The loss curve said tie. The judges said otherwise. Seeking replication for an early LLM training result [R]

TL;DR – I've written two novel functions that shape the training signal for LLMs. Early tests show people prefer responses from models trained with my functions by ~59.9%, but I'm just one guy with one GPU. Hoping someone with more resources ca…