capability
Local Llm agents
This page lists every AI agent in the MeshKore directory tagged with the Local Llm capability. Agents are sourced from public platforms (GitHub, Hugging Face, npm, PyPI, awesome-list curations, and direct submissions), normalized by the MeshKore worker, and ranked by GitHub stars. Each card links to the agent's profile with details on capabilities, framework, language, freshness, and source attribution.
26 agents in this capability · ranked by popularity
Top 26 Local Llm agents
gguf llama 1b text-generation local-llm llama-cpp
Self-learning LLM runtime — TurboQuant KV-cache (6-8x compression), SONA adaptive learning, FlashAttention…
Local CLI - OpenAI-Compatible Local CLI Coding Agent
Hanseol - OpenAI-Compatible Coding Agent
Hanseol - OpenAI-Compatible Coding Agent
Open-source multi-agent coding tool for your terminal. Powered by Ollama.
An agentic coding tool powered by Ollama - local, private, and fast
Run Claude Code locally on the Bonsai 8B 1-bit MLX model.
MCP executor for Claude Code or Codex that offloads repetitive coding work to cheaper local or flat-rate…
Framework for efficient local LLM interaction
39% faster TTFT, 67% less KV cache, zero config — autotune optimises local LLMs on Ollama, LM Studio, and MLX
Mask sensitive data in documents using a local OpenAI-compatible LLM
A unified interface for multiple LLM providers with image generation, speech-to-text, and function calling…
Generate and update agent config files from LM Studio models for VS Code Copilot, OpenCode, Pi, and Codex.
100% local RAG for Obsidian, Zotero, and Claude Code — LightRAG + Ollama + MCP
MCP server for hot-swapping llama.cpp models in Claude Code sessions
Ask your codebase questions using Ollama and Mnemosyne -- zero-config local code search
Multi-agent coding system powered by local LLMs
Validate structured outputs from LLMs with Ollama and automatic retries.
Bridge API service connecting Ollama with Model Context Protocol (MCP) servers
The Fastest RAG Audit - Generate QA datasets & evaluate RAG systems in Colab, Jupyter, or CLI. Privacy-first…
A sequence-based LLM orchestration framework
Multilingual Parallel Translation Platform with Reflection-based Improvement using Local LLMs
Find the best LLM that runs on your hardware
Ollama inference provider for the NucleusIQ AI agent framework (official ollama Python SDK).
One command launcher for running OpenCode with a local llama.cpp model.