The Workflow
1. Get data → Datasets (load_dataset, stream)
2. Tokenize → Transformers (AutoTokenizer)
3. Load model → Transformers (AutoModel) + PEFT (LoRA config)
4. Train → TRL (SFTTrainer/DPOTrainer) + Accelerate
5. Push to Hub → model.push_to_hub("my-model")
6. Inference → Transformers (pipeline) or vLLM
The Hub as Glue
Every library speaks the Hub's language. Models, datasets, and tokenizers are identified by a org/model-name string. Push and pull is one line. Version control (git-lfs) is built in. The Hub is what makes the libraries a platform, not just a collection of tools.
The entire stack is Apache 2.0 licensed — free to use commercially, modify, and redistribute. This is not true of all model weights (check individual model licenses), but the libraries themselves are fully open.