Live stream set for 2025-07-02 at 14:00:00 Eastern
Ask questions in the live chat about any programming or lifestyle topic.
This livestream will be on YouTube or you can watch below.
🔍 Reviewing Phi-3 LLM on Linux Using Alpaca & Ollama
Posted on June 27, 2025 • by [Your Name]
In this article, I’ll walk you through my experience using Phi-3, a lightweight large language model (LLM) developed by Microsoft, running locally on Linux via Alpaca, a user-friendly front-end for Ollama packaged conveniently as a Flatpak.
Whether you’re exploring on-device AI, seeking open-source alternatives to cloud-based LLMs, or want an affordable and private way to run AI locally, Phi-3 is a great place to start. Let’s dive into the setup, performance, and practical use cases of Phi-3 through Alpaca.
💡 What is Phi-3?
Phi-3 is part of Microsoft’s compact LLM family optimized for speed and memory efficiency. Despite its small size (ranging from 1.3B to 7B parameters), it’s surprisingly capable in reasoning, summarization, and Q&A tasks.
Phi-3 models are released under an MIT license, which makes them ideal for developers, hobbyists, educators, and small businesses looking for unrestricted usage.
🐪 Why Use Alpaca (Ollama) on Linux?
Alpaca is a modern GUI client for Ollama, a framework for running LLMs locally with one command. The Flatpak version makes it easy to install and sandbox on any Linux distribution without dependency conflicts.
🔧 Installation Steps
- Install Alpaca via Flatpak:
flatpak install flathub com.jeffser.Alpaca
- Launch Alpaca and select Phi-3 model:
- From Alpaca’s UI, you can configure temperature, and other model parameters with ease.
📷 Screenshots






▶️ Live Demo Screencast
Watch my real-time demo of Phi-3 running inside Alpaca on Linux:
Results:
Who is the mayor of Toronto?
Produced inaccurate outdated answer to Olivia Chow as the mayor of Toronto.
I need a PHP code snippet to connect to a MySQL database.
Produced accurate syntax PHP code snippet to connect to a MySQL database.
I need a 1080p screenshot of the gnome desktop environment.
Produced elaborate answer to generate a 1080p screenshot of Gnome desktop environment because it is a text-based AI lacking ability.
I need a kotlin code snippet to open the camera using Camera2 API and place the camera view on a TextureView.
Produced incomplete Kotlin code snippet to open the camera using Camera2 API.
I need a blender blend file for fire animation.
Produced elaborate answer to generate a fire animation, but not a Blender Blend file because it is a text-based AI lacking ability.
💼 Hire Me: LLM Installation & AI Tutoring
Want to run LLMs like Phi-3 on your own machine? Need help understanding how AI works or integrating it into your workflow?
I offer:
- ✅ One-on-one tutoring on local AI and prompt engineering
- ✅ Remote setup and configuration of Ollama, Alpaca, and models
- ✅ Lightweight model recommendations for your hardware
- ✅ Ongoing support and training for your team
👉 Contact me here or via the form to schedule a session!
📚 Conclusion
Running Phi-3 on Linux through Alpaca is an empowering experience. It’s fast, efficient, and respects your privacy. Whether you’re an AI enthusiast or a small business owner, local models like Phi-3 offer real value without the overhead of the cloud.
Let me know in the comments what models you’re experimenting with, or reach out if you’d like help setting one up!