AI / MLGo
bifrost
Fastest enterprise AI gateway (50x faster than LiteLLM) with adaptive load balancer, cluster mode, guardrails, 1000+ models support & <100 µs overhead at 5k RPS.
N/A rating (0 reviews)0 installs3,438 GitHub stars
Fastest enterprise AI gateway (50x faster than LiteLLM) with adaptive load balancer, cluster mode, guardrails, 1000+ models support & <100 µs overhead at 5k RPS.
Compatible with
Claude DesktopCursor
Install
Claude Desktop
Add to your claude_desktop_config.json:
{
"mcpServers": {
"bifrost": {
"command": "npx",
"args": [
"-y",
"@modelcontextprotocol/server-bifrost"
]
}
}
}Config File Location
Mac: ~/Library/Application Support/Claude/claude_desktop_config.json
Windows: %APPDATA%\Claude\claude_desktop_config.json
Linux: ~/.config/claude/claude_desktop_config.json
Some servers require additional setup - check the GitHub README for specific instructions.