Skip to content

Commit 0bb00a4

Browse files
ebsmothersFelipe Mello
authored andcommitted
Small readme, config updates (#2157)
1 parent b8406e7 commit 0bb00a4

File tree

5 files changed

+5
-5
lines changed

5 files changed

+5
-5
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -140,7 +140,7 @@ loss=torchtune.modules.loss.CEWithChunkedOutputLoss \
140140
enable_activation_checkpointing=True \
141141
optimizer_in_bwd=False \
142142
enable_activation_offloading=True \
143-
optimizer._component_=torch.optim.AdamW \
143+
optimizer=torch.optim.AdamW \
144144
tokenizer.max_seq_len=4096 \
145145
gradient_accumulation_steps=1 \
146146
epochs=1 \

recipes/configs/llama3/8B_dora.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ lr_scheduler:
6161
num_warmup_steps: 100
6262

6363
loss:
64-
_component_: torch.nn.CrossEntropyLoss
64+
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss
6565

6666
# Training
6767
epochs: 1

recipes/configs/llama3/8B_dora_single_device.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ lr_scheduler:
6363
num_warmup_steps: 100
6464

6565
loss:
66-
_component_: torch.nn.CrossEntropyLoss
66+
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss
6767

6868
# Training
6969
epochs: 1

recipes/configs/llama3/8B_qdora_single_device.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ lr_scheduler:
6464
num_warmup_steps: 100
6565

6666
loss:
67-
_component_: torch.nn.CrossEntropyLoss
67+
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss
6868

6969
# Training
7070
epochs: 1

recipes/configs/llama3_1/405B_qlora.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ lr_scheduler:
6161
num_warmup_steps: 100
6262

6363
loss:
64-
_component_: torch.nn.CrossEntropyLoss
64+
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss
6565

6666
fsdp:
6767
cpu_offload: False

0 commit comments

Comments
 (0)