Listen

Description

Gemma, a new family of lightweight, state-of-the-art open models built for responsible AI development, is introduced by Google.

"Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models" presents a new method for instruction tuning of Large Language Models (LLMs) called Generalized Instruction Tuning (GLAN).

"MuLan: Multimodal-LLM Agent for Progressive Multi-Object Diffusion" addresses the challenge of generating images of multiple objects with spatial relationships and attribute bindings.

"Instruction-tuned Language Models are Better Knowledge Learners" explores how to update factual knowledge in large language models. 

Contact:  sergi@earkind.com

Timestamps:

00:34 Introduction

01:21 Google DeepMind Releases Gemma

03:28 Andrej Karpathy on Gemma's Tokenizer

04:16 Groq Inference Tokenomics: Speed, But At What Cost?

05:51 Fake sponsor

07:44 Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models

09:38 MuLan: Multimodal-LLM Agent for Progressive Multi-Object Diffusion

11:06 Instruction-tuned Language Models are Better Knowledge Learners

12:58 Outro