Skip to content

Commit 763b810

Browse files
authored
update all toml files to use experimental section (#392)
1 parent e991ae4 commit 763b810

File tree

5 files changed

+15
-5
lines changed

5 files changed

+15
-5
lines changed

train_configs/llama2_13b.toml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,11 +33,13 @@ max_norm = 1.0 # grad norm clipping
3333
steps = 1000
3434
data_parallel_degree = -1
3535
tensor_parallel_degree = 1
36-
pipeline_parallel_degree = 1
3736
fp8_linear = ""
3837
compile = false
3938
dataset = "c4"
4039

40+
[experimental]
41+
pipeline_parallel_degree = 1
42+
4143
[checkpoint]
4244
enable_checkpoint = false
4345
folder = "checkpoint"

train_configs/llama2_70b.toml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,11 +33,13 @@ max_norm = 1.0 # grad norm clipping
3333
steps = 1000
3434
data_parallel_degree = -1
3535
tensor_parallel_degree = 8 # 8-way TP
36-
pipeline_parallel_degree = 1
3736
fp8_linear = ""
3837
compile = false
3938
dataset = "c4"
4039

40+
[experimental]
41+
pipeline_parallel_degree = 1
42+
4143
[checkpoint]
4244
enable_checkpoint = false
4345
folder = "checkpoint"

train_configs/llama2_7b.toml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,11 +32,13 @@ max_norm = 1.0 # grad norm clipping
3232
steps = 1000
3333
data_parallel_degree = -1
3434
tensor_parallel_degree = 1 # dp-only would be sufficient for 7B
35-
pipeline_parallel_degree = 1
3635
fp8_linear = ""
3736
compile = false
3837
dataset = "c4"
3938

39+
[experimental]
40+
pipeline_parallel_degree = 1
41+
4042
[checkpoint]
4143
enable_checkpoint = false
4244
folder = "checkpoint"

train_configs/llama3_70b.toml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,11 +33,13 @@ max_norm = 1.0 # grad norm clipping
3333
steps = 1000
3434
data_parallel_degree = -1
3535
tensor_parallel_degree = 8 # 8-way TP
36-
pipeline_parallel_degree = 1
3736
fp8_linear = ""
3837
compile = false
3938
dataset = "c4"
4039

40+
[experimental]
41+
pipeline_parallel_degree = 1
42+
4143
[checkpoint]
4244
enable_checkpoint = false
4345
folder = "checkpoint"

train_configs/llama3_8b.toml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,11 +33,13 @@ max_norm = 1.0 # grad norm clipping
3333
steps = 1000
3434
data_parallel_degree = -1
3535
tensor_parallel_degree = 1
36-
pipeline_parallel_degree = 1
3736
fp8_linear = ""
3837
compile = false
3938
dataset = "c4"
4039

40+
[experimental]
41+
pipeline_parallel_degree = 1
42+
4143
[checkpoint]
4244
enable_checkpoint = false
4345
folder = "checkpoint"

0 commit comments

Comments
 (0)