claim
Modern large language models are trained on web-scraped datasets such as CommonCrawl, C4, and The Pile, which contain hundreds of billions to trillions of tokens.

Authors

Sources

Referenced by nodes (1)