AI models keep getting better, but most AI systems still fail in production. Why?
In this episode of High Bit, Brett Gibson sits down with Ghita Houir Alami, cofounder and CEO of ZeroEntropy, to break down the real bottleneck holding AI agents back: retrieval.
Ghita explains why embeddings alone can’t reliably surface the right information, why tools like Slack search feel so frustrating, and how rerankers add a critical second pass that dramatically improves accuracy. She walks through ZeroEntropy’s approach to training rerankers using pairwise comparisons and Elo-style scoring, and why this method generalizes across domains like code, finance, and biology.
The conversation goes deep into:
If you’re building AI agents, search systems, customer support bots, or internal knowledge tools, this episode explains what’s breaking today, and what has to change for AI to work reliably at scale.
(00:00) What changes when retrieval works
(00:39) What ZeroEntropy builds
(01:42) Why retrieval became the real problem
(03:12) Why search fails (Slack included)
(05:11) Why embeddings fall short
(07:11) Rerankers: the missing layer
(10:11) Why rerankers matter most
(12:44) Pairwise ranking vs scoring
(13:52) Elo scoring for documents
(16:33) Fast rerankers via distillation
(18:07) Why old training methods break
(21:29) Retrieval for AI agents
(24:20) Recency, memory, personalization
(32:06) What reliable retrieval unlocks
(33:42) What’s next for ZeroEntropy
Follow Ghita and ZeroEntropy for more:X@ghita__ha@ZeroEntropy_AILinkedInhttps://www.linkedin.com/in/ghita-houir-alami/https://www.linkedin.com/company/zeroentropy-inc