wire llms, ui tweaks

This commit is contained in:
2026-04-12 11:32:36 -03:00
parent 4de44baf98
commit 0f122fa8f7
22 changed files with 960 additions and 203 deletions

108
mcp_servers/shared_llm.py Normal file
View File

@@ -0,0 +1,108 @@
"""Shared LLM client for MCP server narrative tools.
Multi-provider support — selected via LLM_PROVIDER env var:
groq (default) — Groq API, OpenAI-compatible. Fast, free tier.
Needs GROQ_API_KEY.
anthropic — Direct Anthropic SDK. Needs ANTHROPIC_API_KEY.
bedrock — AWS Bedrock Converse API. Needs AWS credentials.
openai — Any OpenAI-compatible endpoint.
Set OPENAI_API_KEY and optionally OPENAI_BASE_URL.
Usage:
LLM_PROVIDER=groq GROQ_API_KEY=gsk_... python -m mcp_servers.shared
"""
import os
from typing import Literal
Provider = Literal["groq", "anthropic", "bedrock", "openai"]
def _get_provider() -> Provider:
p = os.getenv("LLM_PROVIDER", "groq").lower()
if p in ("groq", "anthropic", "bedrock", "openai"):
return p
return "groq"
async def generate(system_prompt: str, user_content: str, max_tokens: int = 1024) -> str:
"""Call an LLM and return the text response."""
provider = _get_provider()
if provider == "anthropic":
return await _generate_anthropic(system_prompt, user_content, max_tokens)
elif provider == "bedrock":
return await _generate_bedrock(system_prompt, user_content, max_tokens)
else:
# groq, openai, or any OpenAI-compatible provider
return await _generate_openai_compat(system_prompt, user_content, max_tokens)
async def _generate_openai_compat(
system_prompt: str, user_content: str, max_tokens: int
) -> str:
"""OpenAI-compatible API (Groq, OpenAI, local, etc)."""
import openai
provider = _get_provider()
if provider == "groq":
api_key = os.getenv("GROQ_API_KEY")
base_url = "https://api.groq.com/openai/v1"
model = os.getenv("GROQ_MODEL", "llama-3.3-70b-versatile")
else:
api_key = os.getenv("OPENAI_API_KEY")
base_url = os.getenv("OPENAI_BASE_URL", "https://api.openai.com/v1")
model = os.getenv("OPENAI_MODEL", "gpt-4o")
client = openai.AsyncOpenAI(api_key=api_key, base_url=base_url)
response = await client.chat.completions.create(
model=model,
max_tokens=max_tokens,
temperature=0.7,
messages=[
{"role": "system", "content": system_prompt},
{"role": "user", "content": user_content},
],
)
return response.choices[0].message.content
async def _generate_anthropic(
system_prompt: str, user_content: str, max_tokens: int
) -> str:
import anthropic
client = anthropic.AsyncAnthropic()
response = await client.messages.create(
model=os.getenv("ANTHROPIC_MODEL", "claude-sonnet-4-20250514"),
max_tokens=max_tokens,
temperature=0.7,
system=system_prompt,
messages=[{"role": "user", "content": user_content}],
)
return response.content[0].text
async def _generate_bedrock(
system_prompt: str, user_content: str, max_tokens: int
) -> str:
import json
import boto3
bedrock = boto3.client(
"bedrock-runtime",
region_name=os.getenv("AWS_DEFAULT_REGION", "us-east-1"),
)
response = bedrock.converse(
modelId=os.getenv("BEDROCK_MODEL_ID", "anthropic.claude-sonnet-4-20250514-v1:0"),
system=[{"text": system_prompt}],
messages=[{"role": "user", "content": [{"text": user_content}]}],
inferenceConfig={"maxTokens": max_tokens, "temperature": 0.7},
)
return response["output"]["message"]["content"][0]["text"]