Skip to content

Commit ae6b1f1

Browse files
authored
Starter Task 1: Get learning rate for llm_pte_finetuning example from config file
Differential Revision: D75807517 Pull Request resolved: #11445
1 parent e5b8735 commit ae6b1f1

File tree

6 files changed

+11
-5
lines changed

6 files changed

+11
-5
lines changed

examples/llm_pte_finetuning/llama3_config.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,8 @@ dataset:
2424
seed: null
2525
shuffle: True
2626

27+
learning_rate: 5e-3
28+
2729
checkpointer:
2830
_component_: torchtune.training.FullModelHFCheckpointer
2931
checkpoint_dir: /tmp/Llama-3.2-1B-Instruct/

examples/llm_pte_finetuning/phi3_alpaca_code_config.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,8 @@ batch_size: 1
1313
loss:
1414
_component_: torch.nn.CrossEntropyLoss
1515

16+
learning_rate: 5e-3
17+
1618
model:
1719
_component_: torchtune.models.phi3.lora_phi3_mini
1820
lora_attn_modules: ['q_proj', 'v_proj']

examples/llm_pte_finetuning/phi3_config.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,8 @@ batch_size: 1
1212
loss:
1313
_component_: torch.nn.CrossEntropyLoss
1414

15+
learning_rate: 5e-3
16+
1517
model:
1618
_component_: torchtune.models.phi3.lora_phi3_mini
1719
lora_attn_modules: ['q_proj', 'v_proj']

examples/llm_pte_finetuning/qwen_05b_config.yaml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,16 @@ batch_size: 1
1313
loss:
1414
_component_: torch.nn.CrossEntropyLoss
1515

16+
learning_rate: 5e-3
17+
1618
model:
1719
_component_: torchtune.models.qwen2.lora_qwen2_0_5b
1820
lora_attn_modules: ['q_proj', 'k_proj', 'v_proj']
1921
apply_lora_to_mlp: False
2022
lora_rank: 32
2123
lora_alpha: 64
24+
# lr parameter is not supported by lora_qwen2_0_5b function
25+
# lr: 5e-3
2226

2327
checkpointer:
2428
_component_: torchtune.training.FullModelHFCheckpointer

examples/llm_pte_finetuning/runner.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ def main() -> None:
8484
# params run from [param_start, outputs_end]
8585
grad_start = et_mod.run_method("__et_training_gradients_index_forward", [])[0]
8686
param_start = et_mod.run_method("__et_training_parameters_index_forward", [])[0]
87-
learning_rate = 5e-3
87+
learning_rate = cfg.learning_rate
8888
f.seek(0)
8989
losses = []
9090
for i, batch in tqdm(enumerate(train_dataloader), total=num_training_steps):

examples/llm_pte_finetuning/training_lib.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -106,10 +106,6 @@ def eval_model(
106106
token_size = tokens.shape[1]
107107
labels_size = labels.shape[1]
108108

109-
tokens, labels = batch["tokens"], batch["labels"]
110-
token_size = tokens.shape[1]
111-
labels_size = labels.shape[1]
112-
113109
# Fixed length for now. We need to resize as the input shapes
114110
# should be the same passed as examples to the export function.
115111
if token_size > max_seq_len:

0 commit comments

Comments
 (0)