Is ChatGPT right?
Is ChatGPT right?
Key takeaway: Size isn’t everything.
Alignment > Scaling.
By fine-tuning with human feedback, InstructGPT shows we can get better, safer AI without endlessly chasing bigger models.
Linked to Paper 👉 https://buff.ly/3Z2e0v3
Key takeaway: Size isn’t everything.
Alignment > Scaling.
By fine-tuning with human feedback, InstructGPT shows we can get better, safer AI without endlessly chasing bigger models.
Linked to Paper 👉 https://buff.ly/3Z2e0v3
Imagine if AI could actually do what you asked, instead of hallucinating random facts or being toxic.
That’s the idea behind InstructGPT—OpenAI’s game-changing 2022 paper. It’s not just smarter; it listens. 🧵👇
Imagine if AI could actually do what you asked, instead of hallucinating random facts or being toxic.
That’s the idea behind InstructGPT—OpenAI’s game-changing 2022 paper. It’s not just smarter; it listens. 🧵👇
In 2018, OpenAI released a bombshell paper: "Improving Language Understanding by Generative Pre-Training".
It didn’t just change AI—it gave us the roadmap for GPT models we see today. 💡✨
What’s it about? Let’s break it down in 10 tweets that are smarter than a fine-tuned Transformer. 🧵👇
In 2018, OpenAI released a bombshell paper: "Improving Language Understanding by Generative Pre-Training".
It didn’t just change AI—it gave us the roadmap for GPT models we see today. 💡✨
What’s it about? Let’s break it down in 10 tweets that are smarter than a fine-tuned Transformer. 🧵👇
Hit me up via DM or reply!
Hit me up via DM or reply!
Meanwhile, think of Transformers as the Netflix of AI: efficient, addictive, and totally changing the landscape. 🍿🤖
Meanwhile, think of Transformers as the Netflix of AI: efficient, addictive, and totally changing the landscape. 🍿🤖
- 2017 (Transformers): ~$800.
- 2019 (BERT): ~$1,344.
- 2020 (GPT-3): ~$153,600.
- 2023 (GPT-4): ~$1.5M+.
Transformers scaled up but kept costs **lower per token** thanks to parallelism & GPU/TPU advancements.
- 2017 (Transformers): ~$800.
- 2019 (BERT): ~$1,344.
- 2020 (GPT-3): ~$153,600.
- 2023 (GPT-4): ~$1.5M+.
Transformers scaled up but kept costs **lower per token** thanks to parallelism & GPU/TPU advancements.
Reed Hastings wants Netflix to "win" your attention. 🎥🛌
In AI, attention means helping models "win" by focusing on the most important data. 🧠✨
"Attention is All You Need" didn’t just reshape AI—it reshaped how we think about focus.
Reed Hastings wants Netflix to "win" your attention. 🎥🛌
In AI, attention means helping models "win" by focusing on the most important data. 🧠✨
"Attention is All You Need" didn’t just reshape AI—it reshaped how we think about focus.
LLMs: Generate text (ChatGPT 📝).
Image GenAI: Create images (MidJourney 🎨).
MidJourney is GenAI but NOT an LLM.
LLMs are a subset of GenAI, and GenAI is a subset of DL. It’s a hierarchy!
(picture from Sebastian Raschka's Build A Large Language Model from scratch book.)
LLMs: Generate text (ChatGPT 📝).
Image GenAI: Create images (MidJourney 🎨).
MidJourney is GenAI but NOT an LLM.
LLMs are a subset of GenAI, and GenAI is a subset of DL. It’s a hierarchy!
(picture from Sebastian Raschka's Build A Large Language Model from scratch book.)
- implementing the LLM architecture and data preparation process
- pre-training an LLM to create a foundation model
- fine tuning the foundation model to become a personal assistant or text classifier
- implementing the LLM architecture and data preparation process
- pre-training an LLM to create a foundation model
- fine tuning the foundation model to become a personal assistant or text classifier