procedure
The LLM training process consists of two primary stages: (1) Pre-Training, a massive-scale, self-supervised process where the model optimizes a next-token prediction objective to acquire linguistic knowledge and reasoning abilities; and (2) Supervised Fine-Tuning (SFT), where the pre-trained model is trained on a smaller, high-quality dataset of labeled instruction-response pairs to adapt to human intent.

Authors

Sources

Referenced by nodes (3)