🚀 Desktop AI

v1.2.1

RUN GGUF


Run LLMs locally with complete privacy. No internet required.

Scroll to explore
✨ Features

Everything You Need

A powerful desktop AI chat application packed with premium features

🖥️

Multiple Themes

Sleek UI with Mocha, Gruvbox, Synthwave, Matrix & Monokai themes

🔒

Complete Privacy

No internet required — everything runs locally on your machine

GPU Acceleration

Blazing speed with GPU support for faster inference

📱

Remote QR Access

Access from your phone on the same network via QR Code

🧠

Thinking Levels

Configurable reasoning depth: None, Low, Medium, High

🔄

Thinking Parser

Advanced streaming parser with state machine for reasoning tags

🎨

Typewriter Animation

Animated welcome screen with beautiful typewriter effect

🔊

Relax Sounds

Background ambient sounds — Rain, White Noise & more

📤

Export Chats

Export conversations in JSONL, MD, TXT, ZIP formats

🎭

Persona Presets

Pre-configured AI personalities for different use cases

📎

File Attachments

PDF, Images, and Text file support built-in

🗣️

Text-to-Speech

Auto-read AI responses aloud with TTS integration

📱

Phone Mode

Compact phone-style window for minimal footprint

🌐

Dual Engine

Switch between llama-server and llama-cli engines

⚙️ How It Works

Simple & Powerful

Powered by llama.cpp — running GGUF models directly on your hardware

01

Load Model

Select a .gguf file from your device. Supports Qwen, Gemma, DeepSeek, Llama and more.

02

Start Engine

llama-server hosts the model on localhost:11434 with full GPU/CPU optimization.

03

Chat

Interactive interface sends and receives responses via real-time Streaming technology.

04

Remote Access

Separate HTTP server enables phone access via QR Code on your local network.

🆕 What's New

Latest Additions

Remote QR Code

Access the app from your phone on the same network

Streaming Parser

Professional state machine algorithm preventing content leak

Engine Selector

Switch between llama-server and llama-cli

Thinking Mode Toggle

Show/hide Thinking dropdown based on settings

Cross-platform Support

Windows, macOS, Linux support for llama-server

Auto-updater

Automatic application updates

XSS Protection

Security with rehype-sanitize

Electron 33+

Latest version with security patches

🛠️ Tech Stack

Built With Modern Tech

⚛️

React 18

Frontend Framework

🎨

TailwindCSS

Styling

🎬

Framer Motion

Animations

Electron 33

Desktop Runtime

🦙

llama.cpp

AI Engine

📝

react-markdown

Markdown Rendering

📦 Installation

Get Started

Clone, install, and run in minutes

Terminal
# Clone the repository
git clone https://github.com/YASSER-27/RunGguf.git
cd RunGguf

# Install dependencies
npm install

# Run in development mode
npm run dev

# Build production version
npm run package
👤 Developer

Meet the Creator

Yasser-27

@yasser-27

Developer

Yasser-27

This project was built entirely solo from scratch. Special thanks to the llama.cpp team for the amazing engine that makes running GGUF models possible on regular devices.