In this illuminating discussion, hosts JC Bonilla and Ardis Kadiu break down the four fundamental ways AI models become smarter: pre-training (long-term memory), context/prompting (short-term memory), real-time reasoning (inference-time processing), and fine-tuning (specialized learning). Using real-world examples from Bloomberg GPT and Apple's strategy, they explain why bigger models aren't always better and how companies can achieve remarkable results by intelligently combining these different approaches to model intelligence. Kadiu provides a masterclass in understanding AI model development, challenging common assumptions about specialized models while explaining why current AI capabilities are sufficient for most applications over the next 4-5 years.
In this illuminating discussion, hosts JC Bonilla and Ardis Kadiu break down the four fundamental ways AI models become smarter: pre-training (long-term memory), context/prompting (short-term memory), real-time reasoning (inference-time processing), and fine-tuning (specialized learning). Using real-world examples from Bloomberg GPT and Apple's strategy, they explain why bigger models aren't always better and how companies can achieve remarkable results by intelligently combining these different approaches to model intelligence. Kadiu provides a masterclass in understanding AI model development, challenging common assumptions about specialized models while explaining why current AI capabilities are sufficient for most applications over the next 4-5 years.
Post-Thanksgiving Welcome and Updates (00:00:07)
Understanding Model Intelligence: The Four Paths (00:29:06)
Pre-training Deep Dive (00:31:07)
Context and Prompting Insights (00:32:44)
Real-time Reasoning Capabilities (00:34:06)
Fine-tuning and Specialization (00:36:16)
Practical Applications and Cost Considerations (00:42:26)
Industry Examples and Case Studies (00:47:20)
Looking Forward: The Next 4-5 Years (00:49:13)