Transformers handles model loading:
from transformers import AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf")
It supports thousands of models with consistent APIs. Model weights download automatically. Tokenizers, configs, and generation all included. This is your starting point for any fine-tuning project.