In this episode, I walk you through how to run Microsoft’s Phi-3, a small but powerful open-source language model, entirely offline on Linux. Using Alpaca (a Flatpak GUI) with Ollama, we show how easy it is to load Phi-3 locally — no API keys, no subscriptions, just fast and private AI.
Whether you're a developer, AI enthusiast, or curious about open-source tools, this setup is perfect for getting started with local LLMs.
🔧 Tools featured:
Ollama (https://ollama.com)
Alpaca GUI (Flatpak)
Phi-3 Mini model (by Microsoft)
📺 Full YouTube tutorial: https://youtu.be/JMZ5llbvZQ8
💼 Need help? I offer setup & AI tutoring: https://ojamboservices.com/contact
#LocalAI #Phi3 #Ollama #Linux #Flatpak #OpenSource #ArtificialIntelligence #AIPodcast #TechTools #LLM