Overview

OLLamaLLMService provides access to locally-run Ollama models through an OpenAI-compatible interface. It inherits from BaseOpenAILLMService and allows you to run various open-source models locally while maintaining compatibility with OpenAI’s API format for privacy and cost control.

Installation

To use Ollama services, you need to install both Ollama and the Pipecat dependency:
  1. Install Ollama on your system from ollama.com/download
  2. Install Pipecat dependency:
pip install "pipecat-ai[ollama]"
  1. Pull a model (first time only):
ollama pull llama2

Prerequisites

Ollama Local Setup

Before using Ollama LLM services, you need:
  1. Ollama Installation: Download and install Ollama from ollama.com
  2. Model Selection: Pull your desired models (llama2, mistral, codellama, etc.)
  3. Local Service: Ensure Ollama service is running (default port 11434)
  4. Hardware: Sufficient RAM and storage for your chosen models

Configuration

  • No API Keys Required: Ollama runs entirely locally
  • Model Management: Use ollama pull <model> to download models
  • Service URL: Default is http://localhost:11434 (configurable)
Ollama runs as a local service on port 11434. No API key required for complete privacy!