"""LLM factory — Bedrock Converse API (production) or direct Anthropic SDK (local dev).""" import os def get_agent_llm(): """Returns a LangChain chat model for agent orchestration.""" if os.getenv("USE_BEDROCK", "").lower() == "true": from langchain_aws import ChatBedrockConverse return ChatBedrockConverse( model=os.getenv("BEDROCK_MODEL_ID", "anthropic.claude-sonnet-4-20250514-v1:0"), region_name=os.getenv("AWS_DEFAULT_REGION", "us-east-1"), temperature=0.3, max_tokens=4096, ) else: from langchain_anthropic import ChatAnthropic return ChatAnthropic( model=os.getenv("ANTHROPIC_MODEL", "claude-sonnet-4-20250514"), temperature=0.3, max_tokens=4096, )