Build your LLM training/finetuning/instruction-tuning/eval pipeline in any configuration, in <30 lines of code.
Including
Build you custom architecture/dataset, or download easily from Huggingface.
Train in single/multi GPU settings in DDP/FSDP settings.
Finetune (Q)LoRA adapters for low-resource settings.
Automatic setup of best hyperparameters (including automatic selection of largest possible batch size, optimal learning rate, etc.)
Inbuilt model utilities. Including hyperparameter schedulers, regular model checkpointing based on best performance, W&B logging.
Evaluate model on any benchmarks with just one line of code.