Skip to content

Commit 49eac2f

Browse files
committed
Fix #726 grad checkpointing + dynamo work in newer torch, but w/ ddp, ddp optimizer must be disabled
1 parent 82d7496 commit 49eac2f

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

src/open_clip_train/main.py

+6
Original file line numberDiff line numberDiff line change
@@ -422,6 +422,12 @@ def main(args):
422422
original_model = model
423423
if args.torchcompile:
424424
logging.info('Compiling model...')
425+
426+
if args.grad_checkpointing and args.distributed:
427+
logging.info('Disabling DDP dynamo optimizer when grad checkpointing enabled.')
428+
# As of now (~PyTorch 2.4/2.5), compile + checkpointing but DDP optimizer must be disabled
429+
torch._dynamo.config.optimize_ddp = False
430+
425431
model = torch.compile(original_model)
426432

427433
if 'train' not in data:

0 commit comments

Comments
 (0)