Capability · Framework — fine-tuning

Together Fine-Tuning

Together AI offers a managed fine-tuning product for teams who want to customise open-weight models without owning training infrastructure. You upload a JSONL dataset, pick a base model, and Together runs the job on their GPU cluster, returning either downloadable weights or a hosted inference endpoint. Pricing is per-token-trained and typically 10-100x cheaper than DIY on cloud GPUs for small jobs.

Framework facts

Category
fine-tuning
Language
API + Python SDK
License
commercial

Install

pip install together

Quickstart

from together import Together

client = Together(api_key='...')
# Upload dataset
file = client.files.upload(file='train.jsonl', purpose='fine-tune')
# Launch job
job = client.fine_tuning.create(
    training_file=file.id,
    model='meta-llama/Llama-3.1-8B-Instruct',
    n_epochs=3,
    lora=True
)
print(job.id)

Alternatives

  • OpenAI fine-tuning — closed-weight models only
  • Fireworks fine-tuning service
  • Modal / RunPod — DIY with Unsloth or Axolotl
  • Predibase — managed open-source fine-tuning

Frequently asked questions

Hosted fine-tuning or DIY?

Hosted wins for one-off jobs, teams without ML-ops, and when you want to deploy immediately as an endpoint. DIY (Unsloth / Axolotl on Modal or RunPod) wins for frequent training, custom algorithms, or when you need to own the weights outright. Together supports both paths.

Can I download the weights afterward?

Yes for LoRA adapters and typically yes for full weights of open models, subject to the base model's license. Always check — some base models (Llama) have acceptable-use restrictions that carry over.

Sources

  1. Together Fine-Tuning — docs — accessed 2026-04-20