Loading...
Loading...
Use when adding LangChain-based LLM routes or services in Python or Next.js stacks; pair with architect-stack-selector.
npx skill4agent add ajrlewis/ai-skills addon-langchain-llmarchitect-python-uv-fastapi-sqlalchemyarchitect-python-uv-batcharchitect-nextjs-bun-appaddon-rag-ingestion-pipelineaddon-langgraph-agentaddon-llm-judge-evalslangchainconfig/skill_manifest.jsonLLM_PROVIDERopenaianthropicollamaDEFAULT_MODELENABLE_STREAMINGyesnoyesUSE_RAGyesnoMAX_INPUT_TOKENS8000uv add langchain langchain-core langchain-community pydantic-settings tiktokenbun add langchain zoduv add langchain-openai langchain-anthropic langchain-ollama
bun add @langchain/openai @langchain/anthropic @langchain/ollamasrc/{{MODULE_NAME}}/llm/provider.py
src/{{MODULE_NAME}}/llm/chains.py
src/{{MODULE_NAME}}/api/routes/llm.pysrc/lib/llm/langchain.ts
src/lib/llm/chains.ts
src/app/api/llm/chat/route.tsUSE_RAG=yesaddon-llm-judge-evalsconfig/skill_manifest.jsonaddon-langchain-llmaddons"judge_backends": ["langchain"]capabilitiesDEFAULT_MODELJUDGE_MODEL{
"outputText": "string",
"model": "string",
"provider": "string"
}uv run ruff check . || true
uv run mypy src || true
bun run lint || true
rg -n "langchain|outputText|provider" src