Episodes (Page 7)
✨
Meta released Llama 3.1-405B, the largest open source model trained on 15T tokens beating GPT-4 on benchmarks, with 8B and 70B models also receiving significant spec bumps
✨
Clémentine Fourrier leads HuggingFace's OpenLLM Leaderboard, which standardizes model evaluation using high-quality benchmarks with reproducible, centralized scoring to replace lab-specific reports
✨
Reka AI achieved #7 on LMsys leaderboard with only 20 employees and $60M funding, demonstrating that top-tier model performance no longer requires massive teams like OpenAI (600) or Google (950+ co...
✨
Databricks' DBRX and Imbue's 70B model outperform GPT-4o zero-shot on reasoning/coding benchmarks while using 7x less data than Llama 3 70B
✨
Raza Habib of HumanLoop hosts High Agency podcast, flipping the interview dynamic with Shawn Wang to discuss the AI Engineer World's Fair and the relevance of the 'Rise of the AI Engineer' essay on...
✨
James Brady (Head of Engineering) and Adam Wiggins (Cofounder Ink & Switch, Heroku) from Elicit share hiring strategies for AI engineers, defining the role as conventional engineers with LLM and pr...
James Brady
✨
Mike Conover, who led OSS models at Databricks and created Dolly, founded Brightwave as an AI research assistant for investment professionals and announced $6M seed round led by Alessio and Decibel
✨
Discusses code editing benchmarks (WebArena, Sotopia), OpenDevin agent framework, and tensions between academic research and industry implementation of AI systems
Aman Sanger
Graham Neubig
Moritz Hardt
✨
Documents the evolution of context window lengths from 84k tokens (MPT-7B) to current 1M+ token models, covering the competitive 'Context Extension Campaigns' between frontier labs
✨
Covers ICLR 2024's best papers on image generation, vision models, and transformer architectures featuring insights from Durk Kingma, Christian Szegedy, and Ilya Sutskever
Christian Szegedy
Durk Kingma
Ilya Sutskever
✨
Explores the understudied role of adult content in driving AI innovation, from image generation to fine-tuning techniques developed in niche online communities
Jesse Silver
✨
Introduces WebSim.ai, WorldSim, and simulative AI as next frontier beyond RAG, exploring how models can generate coherent multi-step simulations and alternate worlds
✨
Traces the evolution from prompt engineering hacks to structured output frameworks, with Instructor providing clean SDK solution for reliable AI outputs
✨
Covers Elicit's approach to supervising AI research processes through structured automation and verification methods
✨
Rapid-fire updates on the Four Wars framework with NLW covering Sora, Suno, and GPT-4 class landscape reshaping plus Open Source models (Llama 3, Mistral, Grok)
Harrison Chase
✨
Announces AI Engineer World's Fair (June 25-27) as the largest AI engineer conference of 2024 with 4 tracks daily and 500+ attendees expected
Sam Schillace
✨
David Luan (Adept co-founder, ex-OpenAI/Google) explains why Google failed to capitalize on LLM advances despite early leadership in AI research
✨
Mikey Shulman (Suno CEO) discusses evolution from text-to-speech and voice synthesis to music generation using transformer architectures
✨
January 2024 recap covering top research trends including notable advances in architecture and training methodologies across major labs
✨
Soumith Chintala (Meta AI Engineering Lead, PyTorch creator) discusses open source AI's importance for trustworthiness and accessibility