Niels Rogge
December 2023
Training and deploying open-source LLMs
Overview
The rise of open LLMs
The rise of open LLMs
The rise of open LLMs
Chatbot Arena by LMSys
The rise of open LLMs
Chatbot Arena by LMSys
The rise of open LLMs
Chatbot Arena by LMSys
Mixtral already on par with GPT-3.5, better than Gemini Pro
The rise of open LLMs
Training LLMs
Karpathy, 2023
Training LLMs
=> to get a “base model”
Training LLMs
2. Supervised fine-tuning (SFT)
Training LLMs
2. Supervised fine-tuning (SFT)
Training LLMs
2. Supervised fine-tuning (SFT)
Training LLMs
3. Human preference training
Training LLMs
3. Human preference training
Training LLMs
3. Human preference training
Training LLMs
3. Human preference training
Training LLMs
Hugging Face alignment handbook
Deploying LLMs
Deploying LLMs
Deploying LLMs
Deploying LLMs
Deploying LLMs
Deploying LLMs
Deploying LLMs
Why open-source?
Advantages | Disadvantages |
|
|
Why closed-source?
Advantages | Disadvantages |
|
|
Exciting developments
Expect LLMs to become smaller, more capable and run a lot faster
Exciting developments
Sit back and enjoy the race 🍿
Exciting developments
Exciting developments
Thanks for your
attention!
PS: connect with me!
@NielsRogge