PodcastIntel
Sign in Get Started Free
Latent Space: The AI Engineer Podcast
Latent Space: The AI Engineer Podcast

Llama 2, 3 & 4: Synthetic Data, RLHF, Agents on the path to Open Source AGI

Jul 23, 2024 · 1h 5m
AI Summary
  • Meta released Llama 3.1-405B, the largest open source model trained on 15T tokens beating GPT-4 on benchmarks, with 8B and 70B models also receiving significant spec bumps
  • Synthetic data generation was key to Llama 3 training, with focus on pre-training pipelines, scaling laws, and post-training including RLHF vs instruction tuning approaches
  • Thomas Scialom, who led post-training for Llama 2 and 3, discusses tool calling, evals, and the role of synthetic data in training the largest open source AGI models

More from Latent Space: The AI Engineer Podcast

View all episodes →

Get AI Summaries for Every New Episode

Subscribe to Latent Space: The AI Engineer Podcast and get AI summaries, guest tracking, and email digests delivered automatically.

Sign Up Free →