Introduction
Pre-Training
Step 1: Download and preprocess the internet
Step 2: Tokenization
Step 3: Neural network training
Step 4: Inference
Base model
Post-Training: Supervised Finetuning
Conversations
Hallucinations
Knowledge of Self
Models need tokens to think
Things the model cannot do well
Post-Training: Reinforcement Learning
Reinforcement learning
DeepSeek-R1
AlphaGo
Reinforcement learning from human feedback (RLHF)
Preview of things to come
Keeping track of LLMs
Where to find LLMs
원본 영상 링크: https://www.youtube.com/watch?v=7xTGNNLPyMI&t=11381s
영상의 주제: ChatGPT와 같은 대형 언어 모델(LLM; Large Language Model)이 어떻게 작동하는지에 대한 포괄적인 설명
일반인도 이해할 수 있는 수준으로 설명하는 것이 목표