Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 4 additions & 2 deletions examples/llm_finetune/qwen/qwen3_8b_hellaswag_pp_peft.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,6 +25,7 @@ step_scheduler:
ckpt_every_steps: 200
val_every_steps: 100 # will run every x number of gradient steps
num_epochs: 2
max_steps: 100

dist_env:
backend: nccl
Expand All @@ -38,7 +39,6 @@ rng:
model:
_target_: nemo_automodel.NeMoAutoModelForCausalLM.from_pretrained
pretrained_model_name_or_path: Qwen/Qwen3-8B
cache_dir: /nemo-rl/hf_cache/

peft:
_target_: nemo_automodel.components._peft.lora.PeftConfig
Expand Down Expand Up @@ -93,7 +93,9 @@ validation_dataset:

validation_dataloader:
_target_: torchdata.stateful_dataloader.StatefulDataLoader
collate_fn: nemo_automodel.components.datasets.utils.default_collater
collate_fn:
_target_: nemo_automodel.components.datasets.utils.default_collater
pad_seq_len_divisible: 168 # must match the training batch seq_len to avoid static PP shape mismatch

optimizer:
_target_: torch.optim.Adam
Expand Down
Loading