claim
Training data density in large language models is non-uniform over time because older events have had more time to accumulate commentary, analysis, and cross-referencing compared to recent events.

Authors

Sources

Referenced by nodes (1)