Skip to content

Commit 2616771

Browse files
committed
make sharding_first by default
1 parent 87515b4 commit 2616771

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

paddlenlp/trainer/training_args.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -782,7 +782,7 @@ class TrainingArguments:
782782
"Following options are supported:\n"
783783
"- pp_first. the topo order is dp, pp, sharding, mp \n"
784784
"- sharding_first. the topo order is dp, sharding, pp, mp \n"
785-
"Default is None, for pp_first"
785+
"Default is None, for sharding_first"
786786
)
787787
},
788788
)
@@ -2072,7 +2072,7 @@ def _post_init_parallel_degree(self):
20722072
self.expert_tensor_parallel_degree = -1
20732073

20742074
if self.hybrid_parallel_topo_order is None:
2075-
self.hybrid_parallel_topo_order = "pp_first"
2075+
self.hybrid_parallel_topo_order = "sharding_first"
20762076
assert self.hybrid_parallel_topo_order in ["pp_first", "sharding_first"]
20772077

20782078
if self.use_hybrid_parallel and self.enable_auto_parallel:

0 commit comments

Comments
 (0)