File tree Expand file tree Collapse file tree
examples/configs/recipes/llm/performance Expand file tree Collapse file tree Original file line number Diff line number Diff line change 1- defaults : ../../../grpo_math_1B.yaml
2- grpo :
3- num_prompts_per_step : 64
4- num_generations_per_prompt : 32
1+ defaults : ./grpo-qwen3-32b-4n8g.yaml
52checkpointing :
6- enabled : false
73 checkpoint_dir : results/grpo-qwen3-32b-4n4g
84policy :
9- model_name : Qwen/Qwen3-32B
10- train_micro_batch_size : 1
11- max_total_sequence_length : 4096
12- dtensor_cfg :
13- enabled : false
14- optimizer : null
15- scheduler : null
16- make_sequence_length_divisible_by : ${policy.megatron_cfg.tensor_model_parallel_size}
175 megatron_cfg :
18- enabled : true
19- empty_unused_memory_level : 1
206 tensor_model_parallel_size : 2
21- pipeline_model_parallel_size : 4
22- sequence_parallel : true
23- optimizer :
24- lr : 3.0e-07
25- min_lr : 3.0e-08
26- scheduler :
27- lr_warmup_iters : 2
28- lr_warmup_init : 3.0e-08
297 generation :
308 vllm_cfg :
319 tensor_parallel_size : 2
3210logger :
3311 log_dir : logs/grpo-qwen3-32b-4n4g
34- wandb_enabled : true
35- tensorboard_enabled : true
3612 wandb :
37- project : nemo-rl
3813 name : grpo-qwen3-32b-4n4g
3914cluster :
4015 gpus_per_node : 4
41- num_nodes : 4
42-
You can’t perform that action at this time.
0 commit comments