PocketLLM is a powerful, cross-platform AI chat application built with Flutter that brings the power of large language models to your pocket. What makes PocketLLM unique is its ability to download and run AI models directly on your mobile device or connect to locally hosted models on your computer.
-
📱 Mobile-First LLM Integration
- Download and run AI models directly on your mobile device
- Optimized for mobile performance and battery life
- Secure, private, and offline-capable operation
-
💻 Desktop Integration
- Seamless connection to locally hosted models via LLMStudio or Ollama
- Easy model management and switching
- Cross-device synchronization
-
🚀 Advanced Features
- Real-time chat interface with markdown support
- Token usage tracking and cost estimation
- Chat history management with local storage
- Dynamic model switching
- Clean, modern UI with dark mode support
- Responsive design for all screen sizes
- Optimized versions of popular open-source models
- Quantized models for efficient mobile execution
- Various size options (7B, 13B) for different device capabilities
- Connect to locally hosted models via:
- Ollama
- LLMStudio
- Custom API endpoints
- Android 8.0 or later
- iOS 13.0 or later
- Minimum 4GB RAM recommended
- At least 2GB free storage space
- Windows 10/11
- macOS 10.15 or later
- Linux (major distributions)
- Ollama or LLMStudio installed (for local model hosting)
Download from (Comming Soon):
- Download the latest release for your platform from the releases page
- Install Ollama or LLMStudio if you want to host models locally
- Launch PocketLLM and connect to your local model server
# Clone the repository
git clone https://github.com/yourusername/pocketllm.git
cd pocketllm
# Install dependencies
flutter pub get
# Run the app
flutter run
-
Mobile Setup
- Launch PocketLLM
- Browse available models in the Model Hub
- Download your preferred model
- Start chatting immediately
-
Desktop Setup
- Install and configure Ollama or LLMStudio
- Launch PocketLLM
- Connect to your local model server
- Select your model and begin chatting
We welcome contributions! Please see our Contributing Guidelines for details.
This project is licensed under the MIT License - see the LICENSE file for details.
- Flutter team for the amazing framework
- Ollama project for local model hosting
- LLMStudio for desktop integration
- All open-source model providers
Note: This project is currently in active development and open source. A beta version will be released soon! If you'd like to contribute, feel free to clone the repository and submit a pull request. Screenshots and documentation will be updated as the project progresses.