Overview

LocalAI is an OpenAI-compatible API for local inference, providing a drop-in replacement for OpenAI’s API that runs on your hardware.

Key Features

  • OpenAI Compatible: Drop-in replacement for OpenAI API
  • Multiple Formats: Support for GGML, GGUF model formats
  • Docker Support: Easy deployment via Docker
  • REST API: Standard REST endpoints
  • Cross-Platform: Runs on any system with Docker

Installation

docker run -p 8080:8080 localai/localai:latest

From Source

git clone https://github.com/go-skynet/LocalAI
cd LocalAI
make build

Usage

  1. Start Server: Run LocalAI server
  2. Use OpenAI SDK: Use standard OpenAI Python/Node.js SDKs
  3. Point to Local: Set base_url to http://localhost:8080

Supported Models

  • Llama models
  • GPT-2/J
  • BERT
  • Stable Diffusion (images)
  • And more

0 items under this folder.