Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
V
verl
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ZhangXiaoyun
verl
Commits
65ac1294
Commit
65ac1294
authored
Apr 10, 2025
by
Yaoyu Zhu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
fix bugs in dapo config (no dynamic sampling, no token-level loss)
parent
d58782a4
Hide whitespace changes
Inline
Side-by-side
Showing
12 changed files
with
34 additions
and
22 deletions
+34
-22
recipe/dapo/blockelite/dapo_7b_test.sh
+3
-2
recipe/dapo/blockelite/run_dapo_codev_7b_3.1k.sh
+1
-0
recipe/dapo/dapo_7b_test.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_16k.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_16k_err_l0.2_continuous_reward_threshold.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_16k_err_l0.2_continuous_reward_zero.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_16k_err_l1.0_continuous_reward.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_16k_inaccurate_16kdata_err_l0.2_threshold.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_20k_err_l0.2_r1_continuous_reward.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_3.4k.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_8k.sh
+3
-2
recipe/dapo/run_dapo_codev_7b_8k_terminal.sh
+3
-2
No files found.
recipe/dapo/blockelite/dapo_7b_test.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 1))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
1024
\
data.max_response_length
=
3072
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/blockelite/run_dapo_codev_7b_3.1k.sh
View file @
65ac1294
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/dapo_7b_test.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 1))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
1024
\
data.max_response_length
=
3072
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_16k.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_16k_err_l0.2_continuous_reward_threshold.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_16k_err_l0.2_continuous_reward_zero.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_16k_err_l1.0_continuous_reward.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_16k_inaccurate_16kdata_err_l0.2_threshold.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_20k_err_l0.2_r1_continuous_reward.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_3.4k.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_8k.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -69,6 +69,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
16384
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
recipe/dapo/run_dapo_codev_7b_8k_terminal.sh
View file @
65ac1294
...
...
@@ -18,12 +18,12 @@ overlong_buffer_len=$((1024 * 4))
overlong_penalty_factor
=
1.0
# An early version for DAPO
enable_filter_groups
=
Fals
e
enable_filter_groups
=
Tru
e
gen_prompt_bsz
=
512
# NOTE: no filtering here
train_prompt_bsz
=
512
train_prompt_mini_bsz
=
32
n_resp_per_prompt
=
16
use_token_level_loss
=
Fals
e
use_token_level_loss
=
Tru
e
# Ray
RAY_ADDRESS
=
${
RAY_ADDRESS
:-
"http://localhost:8265"
}
...
...
@@ -77,6 +77,7 @@ python3 -m verl.trainer.main_ppo \
data.val_batch_size
=
512
\
data.max_prompt_length
=
2048
\
data.max_response_length
=
8192
\
algorithm.filter_groups.enable
=
${
enable_filter_groups
}
\
actor_rollout_ref.model.path
=
$MODEL_PATH
\
+actor_rollout_ref.model.override_config.attention_dropout
=
0.
\
+actor_rollout_ref.model.override_config.embd_pdrop
=
0.
\
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment