Key hyperparameters to tune:
Learning rate: Most impactful. Try - values on log scale.
LoRA rank: Start at , try and .
Batch size: Larger is usually better if memory allows.
Epochs: Monitor validation loss to find optimal stopping point.
Use grid search or Bayesian optimization. Track everything in your experiment logger.