RWKV App ✨
March 20, 2026 · View on GitHub
Run private, on-device AI on phones and desktops with RWKV App. A local-first playground for chat, speech, vision, and model experimentation.
RWKV App is a privacy-first AI app for Android, iOS, Windows, macOS, and Linux. It lets you download local models, compare them on real hardware, and prototype AI experiences without depending on the cloud. After a model is loaded, inference stays on your device.
Why RWKV App
- Built for real edge devices: Evaluate local models on phones and desktops instead of relying on cloud-only demos.
- One app, multiple workflows: Chat, text-to-speech, and visual understanding live in one place.
- Fast model iteration: Download and switch models from Hugging Face to compare quality, speed, and hardware fit.
- Privacy first: Keep prompts, outputs, and inference on device after the model is loaded.

✨ Core Features
- 📱 Cross-Platform, Local-First: Run on Android, iOS, Windows, macOS, and Linux with on-device inference.
- 🤖 Flexible Model Switching: Download and compare different models from Hugging Face.
- 💬 AI Chat: Explore fluent multi-turn conversations on real hardware.
- 🔊 Text-to-Speech (TTS): Convert text into natural-sounding speech.
- 🖼️ Visual Understanding: Explore image-based AI use cases.
- 🔌 Optional Local API Access: On desktop, you can expose an OpenAI-compatible local endpoint for tooling and experiments.
- 🌓 Dark Mode: Stay comfortable during long sessions.
🚀 Get Started
- Download RWKV App from the official page or the platform links below.
- Open the app and load a chat model that fits your device.
- Start exploring chat, speech, or vision workflows. On desktop, you can also enable the built-in local API endpoint when you need it.
Downloads
Official Download Page: https://rwkv.halowang.cloud/
| RWKV Chat (with See and Talk) | RWKV Sudoku | RWKV Othello | RWKV Music (Another repo) | |
|---|---|---|---|---|
| Android APK Download Link | Google Play / GitHub Release / huggingface / pgyer | huggingface / pgyer | huggingface / pgyer | pgyer |
| iOS | App Store / testflight | - | testflight | - |
| Windows | GitHub Release / huggingface (zip) / huggingface (installer) / QQ Group / Discord | Microsoft Store | ||
| macOS | GitHub Release / huggingface / QQ Group / Discord | - | ||
| Linux | GitHub Release / huggingface | - | - | - |
Note
In the future, we will integrate all separate features into the RWKV Chat app to provide a unified experience.
First Run
When you first open the app, a model selection panel will appear. Please choose the model weights you want to use based on your needs.
Warning
Devices older than the iPhone 14 may not be able to smoothly run models with 1.5B / 2.9B parameters.
💻 Build From Source
Ensure you have the Flutter development environment set up.
Development requires Flutter 3.41.1+ (stable channel recommended).
- Clone the repository:
# MUST switch to the 'dev' branch
git clone -b dev https://github.com/MollySophia/rwkv_mobile_flutter.git
# Make sure the rwkv_mobile_flutter and RWKV_APP are in the same directory
git clone -b dev https://github.com/RWKV-APP/RWKV_APP.git
cd RWKV_APP
Project layout should look like this:
parent/
├─ rwkv_mobile_flutter/
└─ RWKV_APP/
- Create necessary configuration files:
touch assets/filter.txt;touch .env;
- Install dependencies:
flutter pub get
- (Optional) Install dependencies for the
toolsdirectory:
Doing this avoids the “Errors exist in your project” warning in VS Code and Cursor when you run the app.
cd tools; flutter pub get; cd ..;
- Run the application:
flutter run
Windows ARM64 Debug (QNN)
If you are debugging on Windows ARM64, uncomment the following section in pubspec.yaml:
- path: assets/lib/qnn-windows/
platforms: [windows]
For Windows ARM64 debugging, check out Flutter's master branch instead of the stable branch.
🏗️ Stack
- Flutter: An open-source framework for building cross-platform user interfaces, supporting Android, iOS, Windows, and macOS.
- Dart FFI (Foreign Function Interface): Used for efficient communication between Dart and the C++ inference engine.
- C++ Inference Engine: The core on-device inference engine, built with C++, supporting multiple model formats and hardware acceleration (CPU/GPU/NPU).
- Hugging Face: An open-source community providing models, datasets, and tools; used here as the source for model weights.
🤝 Feedback and Contribution
This is an experimental early-stage version, and your feedback is crucial to us!
- 🐞 Found a bug or issue? Report it here!
- 💡 Have a suggestion? Suggest a feature!
- 🎨 Want to contribute a custom theme? Theme quick start
📄 License
This project is licensed under the Apache License 2.0. See the LICENSE file for details.