"Almost everyone has heard about large language models, and tens of millions of people have tried out OpenAI ChatGPT and Google Bard. However, the intricate architecture and underlying mathematics driving these remarkable systems remain elusive to many.
LLM's are fascinating - so let's grab a drink and find out how these systems are built and dive deep into their inner workings. In the length of time it to enjoy a round of drinks, you'll understand the inner workings of these models. We'll take our first sip of word vectors, enjoy the refreshing taste of the transformer, and drain a glass understanding how these models are trained on phenomenally large quantities of data.
Large language models for your streaming application - explained with a little maths and a lot of pub stories"
1 of 20
Download to read offline
More Related Content
Explaining How Real-Time GenAI Works in a Noisy Pub
15. 15
How to train your LLM
Pretraining (once)
1. PBs of text.
2. 1000s of GPUs.
3. Compress the text into
a neural network,
4. Pay wait
Obtain base model.
Fine-tuning (recurring)
1. 1000s ideal Q&A
responses (human)
2. Finetune base model
on this data wait
3. Obtain assistant model
4. Evaluate, deploy &
monitor
16. 鶲 It is faster to pick than to generate.
16
Reinforcement Learning from Human Feedback
(RLHF)
鐃緒申 鐃緒申
鐃緒申
17. Training
Pre training & 鍖ne tuning
Large language models
Data
A bunch of text
Architecture
The transformer
17