Use your locally running AI models to assist you in your web browsing
-
Updated
Jan 25, 2026 - TypeScript
Use your locally running AI models to assist you in your web browsing
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
A generalized information-seeking agent system with Large Language Models (LLMs).
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Privacy focused AI powered meeting notes using locally hosted Small Language Models
A flexible, AI powered C2 framework built with operators in mind
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
The PyVisionAI Official Repo
MVP of an idea using multiple local LLM models to simulate and play D&D
Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand
Fenix Ai Trading Bot with LangGraph and ollama and multipe providers
A Cli, a webUI, and a MCP server for the Z-Image-Turbo text-to-image generation model (Tongyi-MAI/Z-Image-Turbo base model as well as quantized models)
Unified management and routing for llama.cpp, MLX and vLLM models with web dashboard.
Open-source AI IDE powered by local & cloud LLMs. A privacy-first alternative to Cursor.
Add a description, image, and links to the localllm topic page so that developers can more easily learn about it.
To associate your repository with the localllm topic, visit your repo's landing page and select "manage topics."