#atom

The foundational learning phase where models acquire general language knowledge

Core Idea: Pre-training is the initial, resource-intensive phase of LLM development where models learn language patterns and world knowledge by predicting the next token in massive text datasets.

Key Elements

Process and Methodology

Knowledge Characteristics

Technical Specifications

Limitations

Connections

References

  1. "Language Models are Few-Shot Learners" (GPT-3 paper)
  2. Chinchilla scaling laws research
  3. Anthropic's research on constitutional AI pre-training approaches

#LLM #pre-training #model-development #next-token-prediction


Connections:


Sources: