Skip to content

Commit 569640b

Browse files
authored
[PaddleNLP->PaddleFormers] Remove restrictions on the use of allgather_overlap #10741 (#2811)
1 parent d80b68d commit 569640b

File tree

1 file changed

+1
-6
lines changed

1 file changed

+1
-6
lines changed

paddleformers/trainer/training_args.py

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1603,12 +1603,7 @@ def is_context_parallel_supported():
16031603
assert (
16041604
"split_param" not in sharding_parallel_config
16051605
), "split_param should not be set when enable_stage1_broadcast_overlap."
1606-
use_casual_mask = os.getenv("USE_CASUAL_MASK", "False")
1607-
assert use_casual_mask, "enable_stage1_broadcast_overlap requires USE_CASUAL_MASK=True."
1608-
assert self.logging_steps > 1, (
1609-
"The logging_steps should be greater than 1 for stage1_broadcast_overlap, "
1610-
f"but got logging_steps={self.logging_steps}."
1611-
)
1606+
16121607
if "enable_stage1_allgather_overlap" in sharding_parallel_config:
16131608
assert (
16141609
ShardingOption.SHARD_OP in self.sharding

0 commit comments

Comments
 (0)