File tree Expand file tree Collapse file tree 5 files changed +5
-5
lines changed Expand file tree Collapse file tree 5 files changed +5
-5
lines changed Original file line number Diff line number Diff line change @@ -140,7 +140,7 @@ loss=torchtune.modules.loss.CEWithChunkedOutputLoss \
140140enable_activation_checkpointing=True \
141141optimizer_in_bwd=False \
142142enable_activation_offloading=True \
143- optimizer._component_ =torch.optim.AdamW \
143+ optimizer=torch.optim.AdamW \
144144tokenizer.max_seq_len=4096 \
145145gradient_accumulation_steps=1 \
146146epochs=1 \
Original file line number Diff line number Diff line change @@ -61,7 +61,7 @@ lr_scheduler:
6161 num_warmup_steps : 100
6262
6363loss :
64- _component_ : torch.nn.CrossEntropyLoss
64+ _component_ : torchtune.modules.loss.CEWithChunkedOutputLoss
6565
6666# Training
6767epochs : 1
Original file line number Diff line number Diff line change @@ -63,7 +63,7 @@ lr_scheduler:
6363 num_warmup_steps : 100
6464
6565loss :
66- _component_ : torch.nn.CrossEntropyLoss
66+ _component_ : torchtune.modules.loss.CEWithChunkedOutputLoss
6767
6868# Training
6969epochs : 1
Original file line number Diff line number Diff line change @@ -64,7 +64,7 @@ lr_scheduler:
6464 num_warmup_steps : 100
6565
6666loss :
67- _component_ : torch.nn.CrossEntropyLoss
67+ _component_ : torchtune.modules.loss.CEWithChunkedOutputLoss
6868
6969# Training
7070epochs : 1
Original file line number Diff line number Diff line change @@ -61,7 +61,7 @@ lr_scheduler:
6161 num_warmup_steps : 100
6262
6363loss :
64- _component_ : torch.nn.CrossEntropyLoss
64+ _component_ : torchtune.modules.loss.CEWithChunkedOutputLoss
6565
6666fsdp :
6767 cpu_offload : False
You can’t perform that action at this time.
0 commit comments