Download
Local AI Runtime for iOS

Run a private AI server directly on your iPhone.

ai.local runs models on-device and exposes local endpoints compatible with both OpenAI API format and Ollama, so you can keep your existing client integrations while keeping data local.

OpenAI API Compatible Ollama Compatible Offline Friendly
ai.local iPhone app screenshot 1 ai.local iPhone app screenshot 2

Compatibility

One local runtime, two standards.

OpenAI API request format Compatible
Ollama-style endpoints Compatible
Local-first execution Default
Cloud dependency required No

API Examples

Familiar formats, local execution.

OpenAI API Style

Use familiar `/v1` conventions.

POST /v1/chat/completions
{
  "model": "qwen3",
  "messages": [{"role": "user", "content": "Summarize this note"}]
}

Ollama Style

Use Ollama-compatible request flows.

POST /api/chat
{
  "model": "qwen3",
  "messages": [{"role": "user", "content": "Summarize this note"}]
}

Supported Open-Source Models

Run popular models locally on your device.

Llama 3

Meta's powerful language model with excellent reasoning capabilities and broad knowledge base.

Mistral

High-performance model from Mistral AI, optimized for efficiency and quality responses.

Qwen

Alibaba's advanced language model with strong multilingual support and coding abilities.

Gemma

Google's lightweight yet powerful model designed for efficient local execution.

Phi-3

Microsoft's compact but capable model, perfect for mobile devices with limited resources.

Stable LM

Stability AI's language model with balanced performance and efficiency for local use.

How It Works

From install to local AI in three steps.

01

Choose a model

Pick the model that fits your device and performance target.

02

Start ai.local server

Serve local endpoints compatible with OpenAI API and Ollama.

03

Connect your tools

Run private workflows from scripts, apps, and shortcuts.

Where It Fits

Perfect for privacy-focused workflows.

Use Cases

ai.local works well for offline demos, privacy-sensitive testing, and local-first prototypes. Teams using OpenAI API clients or Ollama tooling can keep their integration style.

Requirements

Best experience is on iOS 18.2+ and iPadOS 18.2+ with newer high-performance devices, including iPhone 15 Pro, iPhone 16 series, and iPad Pro models with M1/M2 chips for stable local model execution.