torchgpipe: On-the-fly Pipeline Parallelism for Training Giant Models

On the fly
DOI: 10.48550/arxiv.2004.09910 Publication Date: 2020-01-01
ABSTRACT
We design and implement a ready-to-use library in PyTorch for performing micro-batch pipeline parallelism with checkpointing proposed by GPipe (Huang et al., 2019). In particular, we develop set of components to enable pipeline-parallel gradient computation PyTorch's define-by-run eager execution environment. show that each component is necessary fully benefit from such environment, demonstrate the efficiency applying it various network architectures including AmoebaNet-D U-Net. Our available at https://github.com/kakaobrain/torchgpipe .
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....