Overview
Ollama is an easy-to-use local LLM runner that supports multiple open-source models including Llama 3, Mistral, Phi, CodeLlama, and more.
Key Features
- Simple Installation: Single command installation
- Model Library: Access to hundreds of models via registry
- Cross-Platform: Windows, macOS, Linux support
- REST API: Built-in API for integration
- GPU Acceleration: Automatic GPU detection and utilization
Installation
macOS
brew install ollamaLinux/Windows
Download from ollama.ai
Usage
Pull and Run a Model
# Pull a model
ollama pull llama3
# Run the model
ollama run llama3List Available Models
ollama listPopular Models
- llama3 (8B, 70B) - Meta’s latest Llama model
- mistral (7B) - Fast and capable model
- codellama (7B, 13B, 34B) - Code-focused model
- phi3 (3.8B) - Microsoft’s efficient model
Integration
Ollama provides a REST API at http://localhost:11434 for easy integration with other tools and applications.