examples/qwen_coder/ project bundled with the SDK. It fine-tunes the 30B A3B Qwen Coder model using LoRA or QLoRA adapters.
Prerequisites
- Task app deployed (or hosted) and accessible (
TASK_APP_URLin.env). uvx synth-ai setuphas been run in the repo so.envcontainsSYNTH_API_KEYandENVIRONMENT_API_KEY.- Dataset JSONL prepared in the Synth SFT schema.
 - Access to at least 2 × H200 GPUs (the configs default to this).
 
Minimal command
ft:Qwen/...).
Config highlights
use_qlora, r, alpha, and the learning rate to explore different adapter trade-offs. For smaller runs you can switch to coder_lora_4b.toml or coder_lora_small.toml.
Tips
- Set 
[hyperparameters].lora_rankor[training].max_stepswhen sweeping adapter size or training duration. - Use 
--idempotencyto guard against duplicate job submissions in automation. - Upload evaluation datasets and reference this fine-tuned model in your RL configs to continue iterating.