Loading...
Loading...
Monitor Render services in real-time. Check health, performance metrics, logs, and resource usage. Use when users want to check service status, view metrics, monitor performance, or verify deployments are healthy.
npx skill4agent add render-oss/skills render-monitorlist_services()render --versionRENDER_API_KEYrender whoami -o jsonget_selected_workspace()render workspace current -o jsonNote: MCP tools require the Render MCP server. If unavailable, use the CLI for status and logs; metrics and database queries require MCP.
list_services()https://dashboard.render.com/u/*/settings#api-keys~/.cursor/mcp.json<YOUR_API_KEY>{
"mcpServers": {
"render": {
"url": "https://mcp.render.com/mcp",
"headers": {
"Authorization": "Bearer <YOUR_API_KEY>"
}
}
}
}list_services()https://dashboard.render.com/u/*/settings#api-keys<YOUR_API_KEY>claude mcp add --transport http render https://mcp.render.com/mcp --header "Authorization: Bearer <YOUR_API_KEY>"list_services()https://dashboard.render.com/u/*/settings#api-keysexport RENDER_API_KEY="<YOUR_API_KEY>"codex mcp add render --url https://mcp.render.com/mcp --bearer-token-env-var RENDER_API_KEYlist_services()Set my Render workspace to [WORKSPACE_NAME]# 1. Check service status
list_services()
# 2. Check latest deploy
list_deploys(serviceId: "<service-id>", limit: 1)
# 3. Check for errors
list_logs(resource: ["<service-id>"], level: ["error"], limit: 20)
# 4. Check resource usage
get_metrics(resourceId: "<service-id>", metricTypes: ["cpu_usage", "memory_usage"])
# 5. Check latency
get_metrics(resourceId: "<service-id>", metricTypes: ["http_latency"], httpLatencyQuantile: 0.95)list_services()get_service(serviceId: "<id>")list_deploys(serviceId: "<service-id>", limit: 5)| Status | Meaning |
|---|---|
| Deployment successful |
| Building |
| Build failed |
| Replaced by newer deploy |
list_logs(resource: ["<service-id>"], level: ["error"], limit: 50)list_logs(resource: ["<service-id>"], statusCode: ["500", "502", "503"], limit: 50)get_metrics(
resourceId: "<service-id>",
metricTypes: ["cpu_usage", "memory_usage", "cpu_limit", "memory_limit"]
)| Metric | Healthy | Warning | Critical |
|---|---|---|---|
| CPU | <70% | 70-85% | >85% |
| Memory | <80% | 80-90% | >90% |
get_metrics(
resourceId: "<service-id>",
metricTypes: ["http_latency"],
httpLatencyQuantile: 0.95
)| p95 Latency | Status |
|---|---|
| <200ms | Excellent |
| 200-500ms | Good |
| 500ms-1s | Concerning |
| >1s | Problem |
get_metrics(
resourceId: "<service-id>",
metricTypes: ["http_request_count"]
)get_metrics(
resourceId: "<service-id>",
metricTypes: ["http_latency"],
httpPath: "/api/users"
)list_postgres_instances()
get_postgres(postgresId: "<postgres-id>")get_metrics(resourceId: "<postgres-id>", metricTypes: ["active_connections"])query_render_postgres(
postgresId: "<postgres-id>",
sql: "SELECT state, count(*) FROM pg_stat_activity GROUP BY state"
)query_render_postgres(
postgresId: "<postgres-id>",
sql: "SELECT query, mean_exec_time FROM pg_stat_statements ORDER BY mean_exec_time DESC LIMIT 10"
)list_key_value()
get_key_value(keyValueId: "<kv-id>")list_logs(resource: ["<service-id>"], limit: 100)list_logs(resource: ["<service-id>"], level: ["error"], limit: 50)list_logs(resource: ["<service-id>"], text: ["timeout", "error"], limit: 50)list_logs(
resource: ["<service-id>"],
startTime: "2024-01-15T10:00:00Z",
endTime: "2024-01-15T11:00:00Z"
)render logs -r <service-id> --tail -o text# Services
list_services()
get_service(serviceId: "<id>")
list_deploys(serviceId: "<id>", limit: 5)
# Logs
list_logs(resource: ["<id>"], level: ["error"], limit: 100)
list_logs(resource: ["<id>"], text: ["search"], limit: 50)
# Metrics
get_metrics(resourceId: "<id>", metricTypes: ["cpu_usage", "memory_usage"])
get_metrics(resourceId: "<id>", metricTypes: ["http_latency"], httpLatencyQuantile: 0.95)
get_metrics(resourceId: "<id>", metricTypes: ["http_request_count"])
# Database
list_postgres_instances()
get_postgres(postgresId: "<id>")
query_render_postgres(postgresId: "<id>", sql: "SELECT ...")
get_metrics(resourceId: "<postgres-id>", metricTypes: ["active_connections"])
# Key-Value
list_key_value()
get_key_value(keyValueId: "<id>")# Service status
render services -o json
render services instances <service-id>
# Deployments
render deploys list <service-id> -o json
# Logs
render logs -r <service-id> --tail -o text # Stream logs
render logs -r <service-id> --level error -o json # Error logs
render logs -r <service-id> --type deploy -o json # Build logs
# Database
render psql <database-id> # Connect to PostgreSQL
# SSH for live debugging
render ssh <service-id>| Indicator | Healthy | Warning | Critical |
|---|---|---|---|
| Deploy Status | | | |
| Error Rate | <0.1% | 0.1-1% | >1% |
| p95 Latency | <500ms | 500ms-2s | >2s |
| CPU Usage | <70% | 70-90% | >90% |
| Memory Usage | <80% | 80-95% | >95% |