Live stream set for 2025-12-04 at 14:00:00 Eastern
Ask questions in the live chat about any programming or lifestyle topic.
This livestream will be on YouTube or you can watch below.
Steps to Configure Llama.cpp WebUI with Codellama 7B on Fedora 43
In this tutorial, we will go through the steps to configure the Llama.cpp WebUI with Codellama 7B running on a Linux system with an AMD Instinct Mi60 32GB HBM2 GPU. This guide will help you set up the environment, install the necessary software, and ensure that everything works smoothly for optimal performance. If you’re new to machine learning or working with large models, this post is perfect for you.
System Requirements
Before we begin, ensure that your system meets the following requirements:
- Operating System: Linux (tested on Fedora 43)
- GPU: AMD Instinct Mi60 32GB HBM2 (or similar AMD GPU with ROCm support)
- Memory: Minimum 64GB RAM recommended for smooth operation
- Disk Space: At least 100GB free for storing models and data
- Software:
- Linux Kernel version 5.10 or higher
- ROCm (Radeon Open Compute) for AMD GPU acceleration
- Python 3.8 or higher
Is Codellama 7B Open Source?
Yes, Codellama 7B is open source! It is a powerful language model designed to provide high-performance inference on large datasets. It is based on Llama.cpp, a C++ implementation of the LLaMA model. Codellama 7B is licensed under the MIT License, which allows you to use, modify, and distribute the software freely, with a few restrictions such as not holding the authors liable for any issues that arise.
Installation Steps
1. Update Your Fedora System
First, ensure that your Fedora system is up to date. Open a terminal and run:
sudo dnf update -y
2. Install Llama.cpp from Fedora Repositories
Next, install the llama.cpp package from the Fedora repositories:
sudo dnf install llama-cpp
3. Install ROCm for AMD GPU Support
Since you’re using an AMD Instinct Mi60 GPU, you need to install ROCm (Radeon Open Compute) to enable GPU acceleration. Use the following command:
sudo dnf install rocm
4. Verify ROCm Installation
Once ROCm is installed, verify that your AMD GPU is recognized and working correctly by running the following command:
rocminfo
This should show information about your GPU. If it does not, check the official ROCm documentation for troubleshooting.
5. Download Codellama 7B Model Files
Download the Codellama 7B model files. Make sure you store them in an accessible directory on your system (e.g., /models/codellama/).
6. Running Codellama 7B with Llama-server
Once the models are downloaded, you can run the server with the following command:
llama-server -hf ggml-org/codellama-7b-GGUF --jinja -c 0 --host 127.0.0.1 --port 8033
This will start the server at http://127.0.0.1:8033, where you can access the WebUI.
7. Access the WebUI
Once the server is running, open a browser and navigate to http://127.0.0.1:8033 to interact with Codellama 7B.
Screenshots and Screencast
Here’s where you’ll find a visual walkthrough of setting up codellama-7b-hf-q4_k_m.gguf using Alpaca llama.cpp on your local system:







Results:
Who is the mayor of Toronto?
Produced inaccurate current answer to Olivia Chow as the mayor of Toronto.
I need a PHP code snippet to connect to a MySQL database.
Produced no PHP code snippet to connect to a MySQL database.
I need a 1080p screenshot of the gnome desktop environment.
Inaccurately provided instructions to generate a 1080p screenshot of Gnome desktop environment because it is a text-based AI lacking ability.
I need a kotlin code snippet to open the camera using Camera2 API and place the camera view on a TextureView.
Produced no Kotlin code snippet.
I need a blender blend file for fire animation.
Produce gibberish instead of Blender Blend file for a fire animation because it is a text-based AI lacking ability.
Additional Resources
Books
If you are new to Python programming, check out my book Learning Python for a beginner-friendly approach.
Courses
Take my course Learning Python to get hands-on with Python.
One-on-One Python Tutorials
For personalized Python help, contact me via Contact.
Codellama Installation/Support
Need help with Codellama 7B installation or migration? I offer services at Ojambo Services.
Conclusion
By following these steps, you can easily set up and run Codellama 7B on your Fedora 43 system, with llama-server handling the WebUI and providing easy interaction with the model. No additional custom installations are required beyond Fedora’s official packages for both llama.cpp and ROCm.
Disclosure: Some of the links above are referral (affiliate) links. I may earn a commission if you purchase through them - at no extra cost to you.