DiscoverMarvin's MemosGPipe: Easy Scaling with Micro-Batch Pipeline Parallelism
GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism

GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism

Update: 2024-11-02
Share

Description

This episode breaks down the research paper "GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism," which proposes a new method for training very large neural networks by partitioning the model across multiple accelerators and using a novel batch-splitting pipelining algorithm. This approach allows for the efficient training of larger models than previously possible, achieving almost linear speedup with the number of accelerators.

Audio : (Spotify) https://open.spotify.com/episode/4zXyQKSdiSUFK7HkAi6pxO?si=eWWrNsURSqGtw6Phf4tpJg

Paper: https://arxiv.org/abs/1811.06965


Comments 
loading
00:00
00:00
1.0x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism

GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism

Marvin The Paranoid Android