Skip to content

Commit b79b684

Browse files
authored
Clarify setup of optimizer when using empty_init=True (#19067)
1 parent af852ff commit b79b684

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

docs/source-fabric/advanced/model_init.rst

+4
Original file line numberDiff line numberDiff line change
@@ -75,6 +75,10 @@ When training sharded models with :doc:`FSDP <model_parallel/fsdp>` or DeepSpeed
7575
7676
model = fabric.setup(model) # parameters get sharded and initialized at once
7777
78+
# Make sure to create the optimizer only after the model has been set up
79+
optimizer = torch.optim.Adam(model.parameters())
80+
optimizer = fabric.setup_optimizers(optimizer)
81+
7882
.. note::
7983
Empty-init is experimental and the behavior may change in the future.
8084
For FSDP on PyTorch 2.1+, it is required that all user-defined modules that manage parameters implement a ``reset_parameters()`` method (all PyTorch built-in modules have this too).

0 commit comments

Comments
 (0)