Announcement_3

New blog series on Distributed Training started — covering Tensor Parallelism, Pipeline Parallelism, and Activation Recomputation. Read more.