Listen

Description

Join us for an insightful exploration into the cutting-edge design of today's Large Language Models. Seven years on from the original GPT architecture, have we truly seen groundbreaking changes, or are we simply refining existing foundations? This podcast focuses on the architectural developments that define flagship open models in 2025, moving beyond benchmark performance or training algorithms.

In this episode, we'll unpack the key ingredients contributing to LLM performance, examining how developers are pushing the boundaries of efficiency, memory management, and training stability. Discover the evolution and intricacies of:

We'll compare the structural nuances of leading models such as:

Tune in to understand the intricate design decisions driving the next generation of large language models.