Run local AI models for text, images, text-to-speech, and speech-to-text—all in one open-source tool.
No cloud dependency. No complicated setup. Just install, run, and create.
- Overview
- Why Developers Use AI Runner
- Features
- System Requirements
- Quick Start (Docker)
- Installation Details
- AI Models
- Unit Tests
- Database
- Advanced Features
- Missing or Planned Features
- Contributing
AI Runner is a local-first, open-source application that enables you to run:
- Large Language Models (LLMs) for chat and text generation
- Stable Diffusion for image generation and manipulation
- Text-to-Speech (TTS)
- Speech-to-Text (STT)
Originally created as a GUI-centric AI art and chatbot tool for end users, AI Runner has evolved into a developer-friendly platform. With Docker support, an extension API, and a pure Python codebase, you can integrate AI Runner into your own apps or use it as an all-in-one offline inference engine.
Typical Uses:
- AI prototyping: Quickly test local LLMs and image generation.
- Offline scenarios: Work behind firewalls or without internet.
- Custom UI/UX: Build plugins/extensions for your particular domain.
- End-user tools: Hand off a no-code (GUI) solution for less technical stakeholders.
-
Fast Setup with Docker
No need to configure Python environments manually—just pull and run. AI Runner includes all major dependencies, plus GPU support (with NVIDIA Container Toolkit). -
Local LLM & Stable Diffusion in One
Stop juggling separate repos for text generation and image generation. AI Runner unifies them under one interface. -
Plugin & Extension System
Extend or modify AI Runner’s GUI or back-end with custom plugins. Add new model workflows, custom UI panels, or special logic without forking the entire codebase. -
Python Library
Install from PyPi and import AI Runner directly into your Python project (e.g., a game in Pygame or a PySide6 desktop app). -
Offline / Private Data
Keep data on-premise or behind a firewall—great for enterprise or regulated environments that can’t rely on external cloud inference.
If you find it helpful, please star this repo and share it with others—it helps the project grow and signals demand for local AI solutions.
Below is a high-level list of capabilities in AI Runner:
Feature | Description |
---|---|
LLMs & Communication | |
Voice-based chatbot conversations | Have real-time voice-chat sessions with an LLM (speech-to-text + text-to-speech) |
Text-to-speech (TTS) | Convert text to spoken audio using Espeak or SpeechT5 |
Speech-to-text (STT) | Convert spoken audio to text with Whisper |
Customizable chatbots | Create AI personalities and moods for more engaging conversations |
Retrieval-Augmented Generation | Use local doc or website data to enrich chat responses |
Image Generation | |
Stable Diffusion (1.5, SDXL, Turbo) | Generate images from textual prompts, sketches, or existing images |
Drawing tools & ControlNet | Fine-tune image outputs with extra input or guides |
LoRA & Embeddings | Load LoRA models or textual embeddings for specialized image generation |
Image Manipulation | |
Inpaint & Outpaint | Modify portions of generated images while keeping context |
Image filters | Blur, film grain, pixel art, etc. |
Utility | |
Offline | Everything runs locally, no external API required |
Fast generation | E.g., ~2 seconds on an RTX 2080s for stable diffusion |
Docker-based approach | Simplifies setup & ensures GPU acceleration works out of the box |
Dark mode | Built-in theming (Light / Dark / System) |
NSFW toggles | Enable or disable NSFW detection for images |
Ethical guardrails | Basic guardrails for safe LLM usage (optional) |
Extensions | Build your own feature add-ons via the extension API |
Python Library | pip install airunner and embed it in your own projects |
API Support | Optionally use OpenRouter or other external LLMs |
- OS: Linux or Windows
- CPU: Intel i5 (or equivalent)
- Memory: 16 GB RAM
- GPU: NVIDIA RTX 3060 or better (for Stable Diffusion, TTS/Whisper)
- Network: Broadband (used to download models)
- Storage: 130 GB free (model files can be large)
- OS: Ubuntu 22.04
- CPU: Intel i7 (or equivalent)
- Memory: 30+ GB RAM
- GPU: NVIDIA RTX 4090 or higher
- Storage: 130 GB free
- Network: Needed initially for model downloads
Recommended for most developers—it avoids Python environment headaches and streamlines GPU access.
-
Install NVIDIA Container Toolkit
Follow the official guide to enable GPU passthrough for Docker. -
Get the latest docker image
docker pull ghcr.io/capsize-games/airunner/airunner:dev_latest
-
Clone AI Runner and Run Setup
git clone https://github.com/Capsize-Games/airunner.git cd airunner python3 -m venv venv source venv/bin/activate ./src/airunner/bin/setup.sh
- Choose option 1 (Setup xhost)
- Choose option 2 (Install AI Runner scripts)
-
Start AI Runner
airunner-docker airunner
This starts the GUI with stable diffusion, LLM, TTS/STT, and more.
For detailed steps, see the Installation Wiki.
If you prefer not to use Docker, see the Installation Wiki for more information.
By default, AI Runner installs essential TTS/STT and minimal LLM components.
You must supply additional Stable Diffusion models (e.g., from Hugging Face or Civitai).
Organize them under your local AI Runner data directory:
~/.local/share/airunner
├── art
│ └── models
│ ├── SD 1.5
│ │ ├── lora
│ │ └── embeddings
│ ├── SDXL 1.0
│ │ ├── lora
│ │ └── embeddings
│ └── SDXL Turbo
│ ├── lora
│ └── embeddings
To run all tests:
python -m unittest discover -s src/airunner/tests
Or a single test:
python -m unittest src/airunner/tests/test_prompt_weight_convert.py
AI Runner supports a simple database system. See the Wiki for how to:
- Switch engines (SQLite, etc.)
- Make schema changes
- Run migrations
- Memory Optimization: TF32 Mode, VAE/Attention Slicing, Torch 2.0, sequential CPU offload, ToMe token merging.
- Experimental Integrations: Weather-based chatbot prompts, advanced command-line arguments (
--perform-llm-analysis
,--disable-setup-wizard
, etc.). - Safety & Guardrails: Optional NSFW content detection and adjustable guardrails for LLMs.
- Additional model auto-downloaders
- Automated plugin discovery from community repositories
- Fine-tuning workflow for LLMs
- Desktop packaging (PyInstaller or similar)
We welcome pull requests for new features, bug fixes, or documentation improvements. You can also build and share extensions to expand AI Runner’s functionality. For details, see the Extensions Wiki.
If you find this project useful, please consider giving us a ⭐ on GitHub—it really helps with visibility and encourages further development.
Thanks for checking out AI Runner.
Get started with local AI inference in minutes—no more endless environment setup.
Questions or ideas? Join our Discord or open a GitHub Issue.
Happy building!