together-fine-tuning
Original:🇺🇸 English
Translated
5 scriptsChecked / no sensitive code detected
LoRA, full fine-tuning, DPO preference tuning, VLM training, function-calling tuning, reasoning tuning, and BYOM uploads on Together AI. Reach for it whenever the user wants to adapt a model on custom data rather than only run inference, evaluate outputs, or host an existing model.
2installs
Sourcetogethercomputer/skills
Added on
NPX Install
npx skill4agent add togethercomputer/skills together-fine-tuningTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →Together Fine-Tuning
Overview
Use Together AI fine-tuning when the user needs to adapt a model to their own data or behavior.
Supported workflows in this repo:
- LoRA fine-tuning
- full fine-tuning
- DPO preference tuning
- VLM fine-tuning
- function-calling fine-tuning
- reasoning fine-tuning
- BYOM upload paths
When This Skill Wins
- Train a model on custom instruction or conversational data
- Improve function-calling reliability with supervised examples
- Train on preferences rather than only demonstrations
- Fine-tune multimodal or reasoning-oriented models
- Deploy a fine-tuned output model later through dedicated endpoints
Hand Off To Another Skill
- Use for plain inference without training
together-chat-completions - Use to measure a model before or after tuning
together-evaluations - Use to host the resulting tuned model
together-dedicated-endpoints - Use only when the user needs raw infrastructure rather than managed tuning
together-gpu-clusters
Quick Routing
- Standard LoRA or full fine-tuning
- Start with scripts/finetune_workflow.py
- Read references/data-formats.md
- DPO preference tuning
- Start with scripts/dpo_workflow.py
- Function-calling tuning
- Start with scripts/function_calling_finetune.py
- Reasoning tuning
- Start with scripts/reasoning_finetune.py
- VLM tuning
- Start with scripts/vlm_finetune.py
- Model support and deployment options
- Read references/supported-models.md
- Read references/deployment.md
Workflow
- Choose the tuning method that matches the desired behavior change.
- Validate dataset format before spending tokens on training.
- Upload training data and keep the returned file ID.
- Create the job with explicit method-specific parameters.
- Monitor job state, events, and checkpoints before handing off to deployment.
High-Signal Rules
- Python scripts require the Together v2 SDK (). If the user is on an older version, they must upgrade first:
together>=2.0.0.uv pip install --upgrade "together>=2.0.0" - Prefer LoRA unless the user has a specific reason to pay for full fine-tuning.
- Keep data-format validation close to the upload step so bad files fail early.
- Treat deployment as a separate phase; fine-tuning success does not automatically mean serving success.
- Use the method-specific script instead of overloading one generic workflow for all modes.
- Parameterize dataset paths, model IDs, and suffixes in automation instead of embedding one demo dataset forever.
Resource Map
- Data formats: references/data-formats.md
- Supported models: references/supported-models.md
- Deployment guide: references/deployment.md
- LoRA or full workflow: scripts/finetune_workflow.py
- DPO workflow: scripts/dpo_workflow.py
- Function-calling workflow: scripts/function_calling_finetune.py
- Reasoning workflow: scripts/reasoning_finetune.py
- VLM workflow: scripts/vlm_finetune.py