Multi-GPU Fine-tuning with Unsloth
Learn how to fine-tune LLMs on multiple GPUs and parallelism with Unsloth.
See our new Distributed Data Parallel (DDP) multi-GPU Guide here.
Pipeline / model splitting loading
from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
"unsloth/Llama-3.3-70B-Instruct",
load_in_4bit = True,
device_map = "balanced",
)Last updated
Was this helpful?

