Skip to content

Commit

Permalink
update week10
Browse files Browse the repository at this point in the history
  • Loading branch information
zhisbug committed Mar 11, 2024
1 parent 776adfe commit bbf9198
Show file tree
Hide file tree
Showing 3 changed files with 7 additions and 4 deletions.
11 changes: 7 additions & 4 deletions _modules/week-10.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,12 +6,13 @@ status: Active

Mar 11
: **1**{: .label}
: [Slides]() • [Recording]() • [Scribe Notes](#)
: [Slides](assets/slides/22_ml-system-2.pdf) • [Recording]() • [Scribe Notes](#)
: *Reading:*
* [Alpa: Automating Inter- and Intra-Operator Parallelism for Distributed Deep Learning (required)](https://arxiv.org/pdf/2201.12023.pdf)
* [PGPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism (required)](https://arxiv.org/pdf/1909.08053.pdf)
* [FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness (optional)](https://arxiv.org/pdf/2205.14135.pdf)
* [Efficient Memory Management for Large Language Model Serving with PagedAttention (optional)](https://arxiv.org/pdf/2309.06180.pdf)



Expand All @@ -21,9 +22,10 @@ Mar 13
: [Slides]() • [Recording](#) • [Scribe Notes](#)
: *Reading:*
* [Alpa: Automating Inter- and Intra-Operator Parallelism for Distributed Deep Learning (required)](https://arxiv.org/pdf/2201.12023.pdf)
* [PGPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism (required)](https://arxiv.org/pdf/1909.08053.pdf)
* [FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness (optional)](https://arxiv.org/pdf/2205.14135.pdf)
* [Efficient Memory Management for Large Language Model Serving with PagedAttention (optional)](https://arxiv.org/pdf/2309.06180.pdf)



Expand All @@ -33,7 +35,8 @@ Mar 15
: [Slides](#) • [Recording]() • [Scribe Notes](#)
: *Reading:*
* [Alpa: Automating Inter- and Intra-Operator Parallelism for Distributed Deep Learning (required)](https://arxiv.org/pdf/2201.12023.pdf)
* [PGPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [GPipe: Easy Scaling with Micro-Batch Pipeline Parallelism (required)](https://arxiv.org/pdf/1811.06965.pdf)
* [Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism (required)](https://arxiv.org/pdf/1909.08053.pdf)
* [FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness (optional)](https://arxiv.org/pdf/2205.14135.pdf)
* [Efficient Memory Management for Large Language Model Serving with PagedAttention (optional)](https://arxiv.org/pdf/2309.06180.pdf)

Binary file modified assets/slides/21_ml-system-1.pdf
Binary file not shown.
Binary file added assets/slides/22_ml-system-2.pdf
Binary file not shown.

0 comments on commit bbf9198

Please sign in to comment.