Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
raw
history blame contribute delete
213 Bytes
Distributed training with 🤗 Accelerate
As models get bigger, parallelism has emerged as a strategy for training larger models on limited hardware and accelerating training speed by several orders of magnitude.