Loading...
Loading...
Analyze images using OpenAI's Vision API. Use bash command to execute the vision script like 'bash <base_dir>/scripts/vision.sh <image> <question>'. Can understand image content, objects, text, colors, and answer questions about images.
npx skill4agent add zhayujie/chatgpt-on-wechat openai-image-visionenv_config(action="set", key="OPENAI_API_KEY", value="your-key")env_config(action="set", key="OPENAI_API_BASE", value="your-base-url")<available_skills><base_dir>bash# General pattern (MUST start with bash):
bash "<base_dir>/scripts/vision.sh" "<image_path_or_url>" "<question>" [model]
# DO NOT execute the script directly like this (WRONG):
# "<base_dir>/scripts/vision.sh" ...
# Parameters:
# - image_path_or_url: Local image file path or HTTP(S) URL (required)
# - question: Question to ask about the image (required)
# - model: OpenAI model to use (default: gpt-4.1-mini)
# Options: gpt-4.1-mini, gpt-4.1, gpt-4o-mini, gpt-4-turbobash "<base_dir>/scripts/vision.sh" "/path/to/image.jpg" "What's in this image?"bash "<base_dir>/scripts/vision.sh" "https://example.com/image.jpg" "Describe this image in detail"bash "<base_dir>/scripts/vision.sh" "/path/to/photo.png" "What colors are prominent?" "gpt-4o-mini"bash "<base_dir>/scripts/vision.sh" "/path/to/document.jpg" "Extract all text from this image"bash "<base_dir>/scripts/vision.sh" "image.jpg" "List all objects you can see and describe the overall scene"{
"model": "gpt-4.1-mini",
"content": "The image shows...",
"usage": {
"prompt_tokens": 1234,
"completion_tokens": 567,
"total_tokens": 1801
}
}{
"error": "Error description",
"details": "Additional error information"
}