Skip to content

RunanywhereAI/runanywhere-sdks

RunAnywhere Logo

RunAnywhere

On-device AI for mobile apps.
Run LLMs, speech-to-text, and text-to-speech locally—private, offline, fast.

Download on App Store   Get it on Google Play

GitHub Stars License Discord

Chat    Analytics    Structured Output    Voice AI


See It In Action

On-device tool calling demo

Llama 3.2 3B on iPhone 16 Pro Max
Tool calling + LLM reasoning — 100% on-device

View the code  ·  Full tool calling support coming soon


What is RunAnywhere?

RunAnywhere lets you add AI features to your mobile app that run entirely on-device:

  • LLM Chat — Llama, Mistral, Qwen, SmolLM, and more
  • Speech-to-Text — Whisper-powered transcription
  • Text-to-Speech — Neural voice synthesis
  • Voice Assistant — Full STT → LLM → TTS pipeline

No cloud. No latency. No data leaves the device.


SDKs

Platform Status Installation Documentation
Swift (iOS/macOS) Stable Swift Package Manager docs.runanywhere.ai/swift
Kotlin (Android) Stable Gradle docs.runanywhere.ai/kotlin
React Native Beta npm docs.runanywhere.ai/react-native
Flutter Beta pub.dev docs.runanywhere.ai/flutter

Quick Start

Swift (iOS / macOS)

import RunAnywhere
import LlamaCPPRuntime

// 1. Initialize
LlamaCPP.register()
try RunAnywhere.initialize()

// 2. Load a model
try await RunAnywhere.downloadModel("smollm2-360m")
try await RunAnywhere.loadModel("smollm2-360m")

// 3. Generate
let response = try await RunAnywhere.chat("What is the capital of France?")
print(response) // "Paris is the capital of France."

Install via Swift Package Manager:

https://github.com/RunanywhereAI/runanywhere-sdks

Full documentation → · Source code


Kotlin (Android)

import com.runanywhere.sdk.public.RunAnywhere
import com.runanywhere.sdk.public.extensions.*

// 1. Initialize
LlamaCPP.register()
RunAnywhere.initialize(environment = SDKEnvironment.DEVELOPMENT)

// 2. Load a model
RunAnywhere.downloadModel("smollm2-360m").collect { println("${it.progress * 100}%") }
RunAnywhere.loadLLMModel("smollm2-360m")

// 3. Generate
val response = RunAnywhere.chat("What is the capital of France?")
println(response) // "Paris is the capital of France."

Install via Gradle:

dependencies {
    implementation("com.runanywhere.sdk:runanywhere-kotlin:0.1.4")
    implementation("com.runanywhere.sdk:runanywhere-core-llamacpp:0.1.4")
}

Full documentation → · Source code


React Native

import { RunAnywhere, SDKEnvironment } from '@runanywhere/core';
import { LlamaCPP } from '@runanywhere/llamacpp';

// 1. Initialize
await RunAnywhere.initialize({ environment: SDKEnvironment.Development });
LlamaCPP.register();

// 2. Load a model
await RunAnywhere.downloadModel('smollm2-360m');
await RunAnywhere.loadModel(modelPath);

// 3. Generate
const response = await RunAnywhere.chat('What is the capital of France?');
console.log(response); // "Paris is the capital of France."

Install via npm:

npm install @runanywhere/core @runanywhere/llamacpp

Full documentation → · Source code


Flutter

import 'package:runanywhere/runanywhere.dart';
import 'package:runanywhere_llamacpp/runanywhere_llamacpp.dart';

// 1. Initialize
await RunAnywhere.initialize();
await LlamaCpp.register();

// 2. Load a model
await RunAnywhere.downloadModel('smollm2-360m');
await RunAnywhere.loadModel('smollm2-360m');

// 3. Generate
final response = await RunAnywhere.chat('What is the capital of France?');
print(response); // "Paris is the capital of France."

Install via pub.dev:

dependencies:
  runanywhere: ^0.15.11
  runanywhere_llamacpp: ^0.15.11

Full documentation → · Source code


Sample Apps

Full-featured demo applications demonstrating SDK capabilities:

Platform Source Code Download
iOS examples/ios/RunAnywhereAI App Store
Android examples/android/RunAnywhereAI Google Play
React Native examples/react-native/RunAnywhereAI Build from source
Flutter examples/flutter/RunAnywhereAI Build from source

Playground

Standalone demo projects showcasing what you can build with RunAnywhere:

Project Description Platform
swift-starter-app Privacy-first AI demo — LLM Chat, STT, TTS, and Voice Pipeline iOS (Swift/SwiftUI)
on-device-browser-agent On-device AI browser automation — no cloud, no API keys Chrome Extension

Features

Feature iOS Android React Native Flutter
LLM Text Generation
Streaming
Speech-to-Text
Text-to-Speech
Voice Assistant Pipeline
Model Download + Progress
Structured Output (JSON) 🔜 🔜
Apple Foundation Models

Supported Models

LLM (GGUF format via llama.cpp)

Model Size RAM Required Use Case
SmolLM2 360M ~400MB 500MB Fast, lightweight
Qwen 2.5 0.5B ~500MB 600MB Multilingual
Llama 3.2 1B ~1GB 1.2GB Balanced
Mistral 7B Q4 ~4GB 5GB High quality

Speech-to-Text (Whisper via ONNX)

Model Size Languages
Whisper Tiny ~75MB English
Whisper Base ~150MB Multilingual

Text-to-Speech (Piper via ONNX)

Voice Size Language
Piper US English ~65MB English (US)
Piper British English ~65MB English (UK)

Repository Structure

runanywhere-sdks/
├── sdk/
│   ├── runanywhere-swift/          # iOS/macOS SDK
│   ├── runanywhere-kotlin/         # Android SDK
│   ├── runanywhere-react-native/   # React Native SDK
│   ├── runanywhere-flutter/        # Flutter SDK
│   └── runanywhere-commons/        # Shared C++ core
│
├── examples/
│   ├── ios/RunAnywhereAI/          # iOS sample app
│   ├── android/RunAnywhereAI/      # Android sample app
│   ├── react-native/RunAnywhereAI/ # React Native sample app
│   └── flutter/RunAnywhereAI/      # Flutter sample app
│
├── Playground/
│   ├── swift-starter-app/          # iOS AI playground app
│   └── on-device-browser-agent/    # Chrome browser automation agent
│
└── docs/                           # Documentation

Requirements

Platform Minimum Recommended
iOS 17.0+ 17.0+
macOS 14.0+ 14.0+
Android API 24 (7.0) API 28+
React Native 0.74+ 0.76+
Flutter 3.10+ 3.24+

Memory: 2GB minimum, 4GB+ recommended for larger models


Contributing

We welcome contributions. See our Contributing Guide for details.

# Clone the repo
git clone https://github.com/RunanywhereAI/runanywhere-sdks.git

# Set up a specific SDK (example: Swift)
cd runanywhere-sdks/sdk/runanywhere-swift
./scripts/build-swift.sh --setup

# Run the sample app
cd ../../examples/ios/RunAnywhereAI
open RunAnywhereAI.xcodeproj

Support


License

Apache 2.0 — see LICENSE for details.