We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d80b68d commit 569640bCopy full SHA for 569640b
paddleformers/trainer/training_args.py
@@ -1603,12 +1603,7 @@ def is_context_parallel_supported():
1603
assert (
1604
"split_param" not in sharding_parallel_config
1605
), "split_param should not be set when enable_stage1_broadcast_overlap."
1606
- use_casual_mask = os.getenv("USE_CASUAL_MASK", "False")
1607
- assert use_casual_mask, "enable_stage1_broadcast_overlap requires USE_CASUAL_MASK=True."
1608
- assert self.logging_steps > 1, (
1609
- "The logging_steps should be greater than 1 for stage1_broadcast_overlap, "
1610
- f"but got logging_steps={self.logging_steps}."
1611
- )
+
1612
if "enable_stage1_allgather_overlap" in sharding_parallel_config:
1613
1614
ShardingOption.SHARD_OP in self.sharding
0 commit comments