#inference
15 results found
Lemonade
Local LLM Server with NPU Acceleration
MindBridge MCP Server ⚡ The AI Router for Big Brain Moves
MindBridge is an AI orchestration MCP server that lets any app talk to any LLM — OpenAI, Anthropic, DeepSeek, Ollama, and more — through a single unified API. Route queries, compare models, get second opinions, and build smarter multi-LLM workflows.
Together.ai Mcp
A Node.js Model Context Protocol (MCP) server that exposes Together AI's inference endpoints — chat completions, image generation, vision, and embeddings — as tools callable from Claude Desktop, Cursor, VS Code, and any other MCP-compatible client. I created this MCP due to an issue I was having accessing reasoning models through Together AI. Together AI's largest reasoning models (GLM-5, Qwen3.5-397B, MiniMax M2.5, Kimi K2.5) use a non-standard response format. During chain-of-thought generation, these models write their reasoning trace into choices[0].message.reasoning while leaving choices[0].message.content as an empty string. The final answer only appears in message.content once thinking is complete.
Neo Mcp Logic Analyze
Python MCP server for controlled logic analysis from natural language, with an emphasis on auditable output and teaching-oriented explanations. ## What it does This server accepts short natural-language statements and arguments, then provides structured logic-oriented outputs such as: - controlled formalization into propositional logic; - controlled formalization into a restricted fragment of first-order logic; - ambiguity detection relevant to formalization; - consistency checking; - entailment checking; - simple counterexamples when entailment fails; - natural-language explanations of the formalization process.
Awesome Stars
A curated collection of top-tier penetration testing tools and productivity utilities across multiple domains. Join us to explore, contribute, and enhance your hacking toolkit!
GreenDoc - SUSEAI Demo Lab
A set of bundles to demonstrate SUSE AI capabilities
Shinkai Apps
Shinkai is a two click install AI manager (Local and Remote) that allows you to create AI agents in 5 minutes or less using a simple UI. Agents and tools are exposed as an MCP Server.
Anthropic Model Context Protocol (MCP) Server with Ollama Integration
Model Context Protocol (MCP) server integrated with an external inference service (e.g., Ollama/Gemma3) via middleware.
Sunra.ai Mcp Server
Sunra.ai is a generative media platform built for developers, providing high-performance AI model inference capabilities. It specializes in running diffusion models through production-ready APIs and interactive UI playgrounds.
Inception ICORE Server
Model–Context–Protocol (MCP) Server A modular, extensible Rust-based server providing short-term, long-term, and latent memory services, a chat endpoint backed by a BaseAgent + Sentience DSL, and seamless integration with ChromaDB and LLM services.
Gpu Bridge Mcp Server
Access 47 AI models and 30 services via MCP. LLM, image gen, video, speech, embeddings, reranking, PDF parsing & more. Pay-per-use with x402 (USDC) or API key.
Ownrig - Ai Hardware Compatibility
AI hardware compatibility data for running LLMs locally. Query 50 models, 25 GPUs, 9 ready-to-buy machines, and 663 compatibility entries. Get hardware recommendations, check VRAM requirements, and find buy links.
Qsp Mcp
Named after the Q-signal QSP ("Will you relay?"), qsp-mcp relays tool calls between a local LLM and MCP servers. Any model with function calling capability gains access to the full qso-graph tool ecosystem — 71+ tools across 12 servers — from local weights, not from cloud.