Listen

Description

Explore three breakthroughs in transformer architectures: IBM's Bamba, an attention-state space model overcoming the KV cache bottleneck; Delphi-2M, a generative model predicting human disease trajectories from health records; and the first manually labeled Kashmiri news dataset for fine-tuning LLMs in low-resource settings.