Reduce training costs:
- Use spot instances when possible
- Start with smaller models for experimentation
- Use QLoRA to fit on cheaper GPUs
- Profile and eliminate bottlenecks
- Preempt and resume rather than running continuously
A well-optimized setup can be -x cheaper than naive approaches.