Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions intermediate_source/monarch_distributed_tutorial.rst
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ We rely on a nightly build of Titan for this tutorial, so please ensure that oth
`Install script <https://github.com/meta-pytorch/monarch/blob/main/scripts/install_nightly.py>`_
2. **TorchTitan nightly installed:**
`TorchTitan install instructions <https://github.com/pytorch/torchtitan?tab=readme-ov-file#nightly-builds>`_
3. **A valid Titan model config** and **tokenizer** in your working directory (e.g., ``debug_model.toml`` from `TorchTitan configs <https://github.com/pytorch/torchtitan/blob/main/torchtitan/models/llama3/train_configs/debug_model.toml>`_).
3. **A valid Titan model config** and **tokenizer** in your working directory
4. **SLURM cluster access:**

- Sufficient permissions to reserve nodes and launch jobs.
Expand Down Expand Up @@ -446,4 +446,4 @@ TorchTitan for scalable distributed training.
**Further Reading**

- Monarch also integrates with TorchFT to provide per-step fault-tolerance across replicated workers. You can find a comprehensive `proof of concept <https://github.com/meta-pytorch/torchft/tree/main/examples/monarch>`_ of this integration in the TorchFT repo.
- For an interactive notebook covering similar topics to this tutorial, please consult `this Monarch example <https://github.com/meta-pytorch/monarch/blob/main/examples/slurm_titan.ipynb>`_.
- For an interactive notebook covering similar topics to this tutorial, please consult `this Monarch example <https://github.com/meta-pytorch/monarch/blob/main/examples/slurm_titan.ipynb>`_.