Update to README.md with a reference to the official Reexpress AI MCP… #1751
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
… Server
This adds a reference to the official Reexpress AI MCP Server in README.md.
Description
Server Details
The Reexpress MCP Server is a drop-in solution to add state-of-the-art statistical verification to your complex LLM pipelines, as well as your everyday use of LLMs for search and QA for software development and data science settings. It's the first reliable, statistically robust AI second opinion for your AI workflows.
In addition to providing you (the user) with a principled estimate of confidence in the output given your instructions, Claude (or other tool-calling LLMs) itself can use the verification output to progressively refine its answer, determine if it needs additional outside resources or tools, or has reached an impasse and needs to ask you for further clarification or information.
The current version works on Apple silicon Macs running macOS 15 (Sequoia). The server ensembles external LLM APIs with an on-device PyTorch model and an updatable local dense database, over which it calculates a well-calibrated Similarity-Distance-Magnitude estimator.
Motivation and Context
Given the wide-spread use of LLMs in real-world settings, a principled and robust approach for uncertainty quantification over the output of such models is critical. Here, the estimator is relative to the binary classification decision of whether the response addresses the user's question or instruction. This is useful, for example, to constrain hallucinations (up to a suitable probabilisitic threshold for a given task), and to route conditional branching decisions in complex agent-based pipelines.
How Has This Been Tested?
The MCP server has been thoroughly tested with Claude Sonnet 3.7 on Claude Desktop for macOS, as well as with VSCode Copilot. For typical settings, it is recommended to combine web search (or as applicable, a retrieval system) and extended thinking when calling the main Reexpress tool.
Breaking Changes
Adding the server requires a standard update to the MCP client configuration JSON (as in claude_desktop_config.json). We currently use conda, rather than uv, since the official release of the Faiss dependency is distributed through conda.
Types of changes
Checklist
Additional context
The server has an Apache-2.0 license.