Standard settings that work across frameworks:
- Learning rate: e- for LoRA, e- for full fine-tune
- Batch size: As large as memory allows
- Epochs: - for most datasets
- Warmup: -% of total steps
- Weight decay: -
These are starting points. Your data may need different values.