Overview

Synth-AI supports any Hugging Face-compatible causal LM deployed via vLLM, plus OpenAI-compatible endpoints. Below are common choices validated in examples.

Validated examples

  • Qwen 0.6B / 1.8B (HF, via vLLM)
  • Llama 3.x (selected sizes)
  • Mistral 7B

Considerations

  • Ensure tokenizer path aligns with weights if using custom merges.
  • For RL, configure max_tokens, max_model_len, and batch sizes consistent with GPU memory.