A comprehensive GUI application for managing Ollama models on your local system. This tool allows you to list installed models, remove them, and install new models from a curated list with detailed descriptions.
- List Installed Models: View all locally installed Ollama models with detailed information including size, modification date, and technical details
- Remove Models: Safely remove unwanted models to free up disk space
- Install New Models: Browse and install from a curated list of popular models with descriptions
- Model Information: View detailed information about each model including size, family, and capabilities
- Search Functionality: Search through available models by name, family, or description
- Real-time Status: Get real-time feedback on operations with a status bar
- Python 3.7 or higher
- Ollama installed and running on your system
- Internet connection for installing new models
- Clone or download this repository
- Install the required dependencies:
pip install -r requirements.txt
- Make sure Ollama is running on your system (usually on
http://localhost:11434) - Run the application:
python llama_herder.py
- The left panel shows all currently installed models
- Click on a model to view detailed information
- Use the "Refresh" button to update the list
- Select a model from the installed models list
- Click "Remove Selected Model"
- Confirm the deletion in the dialog
- Use the search box to find models by name, family, or description
- Select a model from the available models list
- Read the description to understand the model's capabilities
- Click "Install Selected Model"
- Wait for the installation to complete
The application includes a curated list of popular models:
- Llama 3.2/3.1: Meta's latest models with excellent performance
- Mistral: Efficient models with good performance-to-size ratio
- Mixtral: Advanced mixture of experts models
- Gemma: Google's open-source models
- Phi-3: Microsoft's compact but capable models
- Qwen2.5: Alibaba's multilingual models
- Code Llama: Specialized for code generation and programming tasks
- Neural Chat: Optimized for conversational AI
- Dolphin: Uncensored models for creative tasks
- OpenChat: Open-source conversational models
- API Integration: Uses Ollama's REST API for all operations
- Threading: Operations run in background threads to keep the UI responsive
- Error Handling: Comprehensive error handling with user-friendly messages
- Cross-platform: Works on Windows, macOS, and Linux
- Make sure Ollama is installed and running
- Check that Ollama is accessible at
http://localhost:11434 - Try restarting the Ollama service
- Check your internet connection
- Ensure you have enough disk space
- Some models are large and may take time to download
- Make sure the model is not currently in use
- Try refreshing the model list first
- Check that you have sufficient permissions
Feel free to submit issues, feature requests, or pull requests to improve this application.
This project is open source and available under the MIT License.