Submit

#inference

15 results found

L

Lemonade

Local LLM Server with NPU Acceleration

M

MindBridge MCP Server ⚡ The AI Router for Big Brain Moves

MindBridge is an AI orchestration MCP server that lets any app talk to any LLM — OpenAI, Anthropic, DeepSeek, Ollama, and more — through a single unified API. Route queries, compare models, get second opinions, and build smarter multi-LLM workflows.

T

Together.ai Mcp

A Node.js Model Context Protocol (MCP) server that exposes Together AI's inference endpoints — chat completions, image generation, vision, and embeddings — as tools callable from Claude Desktop, Cursor, VS Code, and any other MCP-compatible client. I created this MCP due to an issue I was having accessing reasoning models through Together AI. Together AI's largest reasoning models (GLM-5, Qwen3.5-397B, MiniMax M2.5, Kimi K2.5) use a non-standard response format. During chain-of-thought generation, these models write their reasoning trace into choices[0].message.reasoning while leaving choices[0].message.content as an empty string. The final answer only appears in message.content once thinking is complete.

N

Neo Mcp Logic Analyze

Python MCP server for controlled logic analysis from natural language, with an emphasis on auditable output and teaching-oriented explanations. ## What it does This server accepts short natural-language statements and arguments, then provides structured logic-oriented outputs such as: - controlled formalization into propositional logic; - controlled formalization into a restricted fragment of first-order logic; - ambiguity detection relevant to formalization; - consistency checking; - entailment checking; - simple counterexamples when entailment fails; - natural-language explanations of the formalization process.

A

Awesome Stars

A curated collection of top-tier penetration testing tools and productivity utilities across multiple domains. Join us to explore, contribute, and enhance your hacking toolkit!

G

GreenDoc - SUSEAI Demo Lab

A set of bundles to demonstrate SUSE AI capabilities

S

Shinkai Apps

Shinkai is a two click install AI manager (Local and Remote) that allows you to create AI agents in 5 minutes or less using a simple UI. Agents and tools are exposed as an MCP Server.

A

Anthropic Model Context Protocol (MCP) Server with Ollama Integration

Model Context Protocol (MCP) server integrated with an external inference service (e.g., Ollama/Gemma3) via middleware.

S

Sunra.ai Mcp Server

Sunra.ai is a generative media platform built for developers, providing high-performance AI model inference capabilities. It specializes in running diffusion models through production-ready APIs and interactive UI playgrounds.

I

Inception ICORE Server

Model–Context–Protocol (MCP) Server A modular, extensible Rust-based server providing short-term, long-term, and latent memory services, a chat endpoint backed by a BaseAgent + Sentience DSL, and seamless integration with ChromaDB and LLM services.

G

Gpu Bridge Mcp Server

Access 47 AI models and 30 services via MCP. LLM, image gen, video, speech, embeddings, reranking, PDF parsing & more. Pay-per-use with x402 (USDC) or API key.

O

Ownrig - Ai Hardware Compatibility

AI hardware compatibility data for running LLMs locally. Query 50 models, 25 GPUs, 9 ready-to-buy machines, and 663 compatibility entries. Get hardware recommendations, check VRAM requirements, and find buy links.

Q

Qsp Mcp

Named after the Q-signal QSP ("Will you relay?"), qsp-mcp relays tool calls between a local LLM and MCP servers. Any model with function calling capability gains access to the full qso-graph tool ecosystem — 71+ tools across 12 servers — from local weights, not from cloud.

© 2025 MCP.so. All rights reserved.

Build with ShipAny.