cpu-only
Here are 8 public repositories matching this topic...
🦙 chat-o-llama: A lightweight, modern web interface for AI conversations with support for both Ollama and llama.cpp backends. Features persistent conversation management, real-time backend switching, intelligent context compression, and a clean responsive UI.
- 
            Updated
            
Jul 18, 2025  - Python
 
An LLM-based content moderator. Firefox extension to block webpages unrelated to work, based on page title and URL. Local LLMs with Ollama and Langchain to ensure your browsing history never leaves your device, for complete privacy. Google Gemini also supported.
- 
            Updated
            
Dec 12, 2024  - Python
 
Image Classification with On-Device Inference, built with Flutter, AI model runs on mobile cpu
- 
            Updated
            
Jan 29, 2025  - Dart
 
CPU-friendly experience-based reasoning framework combining meta-learning (MAML), state space models (SSM), and memory buffers for fast few-shot adaptation. Pure NumPy implementation for edge devices and low-compute environments.
- 
            Updated
            
Oct 23, 2025  - Python
 
Face Detection service, super fast inference with a nano model
- 
            Updated
            
Jan 26, 2025  - Python
 
A new one shot face swap approach for image and video domains - version tailored to work on CPU
- 
            Updated
            
Aug 20, 2024  - Python
 
Chat-O-Llama is a user-friendly web interface for managing conversations with Ollama, featuring persistent chat history. Easily set up and start your chat sessions with just a few commands. 🐙💻
- 
            Updated
            
Nov 4, 2025  - HTML
 
Improve this page
Add a description, image, and links to the cpu-only topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the cpu-only topic, visit your repo's landing page and select "manage topics."