Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
-
Updated
Mar 25, 2026 - Go
Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
Android native AI inference library, bringing gguf models and stable-diffusion inference on android devices, powered by llama.cpp and stable-diffusion.cpp
A thin cython wrapper around llama.cpp, whisper.cpp and stable-diffusion.cpp
A Front End for stable-diffusion.cpp. Built with Electron and Node.js, I made it to be simple to interact with the stable difussion cpp cli, only tested using the Vulkan binary compiled for Windows.
Examples using the llmedge library
Pre-built stable-diffusion.cpp binaries for Leaxer
Fully Local AI (FLAI) is a self-hosted, privacy-first AI platform that gives you a fully functional personal assistant — chat, voice, image generation and editing, document queries, and camera image analysis — all locally, without sending data to the cloud. It supports multi-user access with request queuing, is GPU‑optimized, and is open source.
an early-stage experimental nanobind wrapper around llama.cpp
Add a description, image, and links to the stable-diffusion-cpp topic page so that developers can more easily learn about it.
To associate your repository with the stable-diffusion-cpp topic, visit your repo's landing page and select "manage topics."