Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
V
verl
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ZhangXiaoyun
verl
Commits
55093874
Unverified
Commit
55093874
authored
Feb 16, 2025
by
HL
Committed by
GitHub
Feb 16, 2025
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
example: fix remove padding flags for gemma example. update v0.2 install docs (#290)
parent
0c32cf78
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
5 additions
and
8 deletions
+5
-8
docs/start/install.rst
+3
-6
examples/ppo_trainer/run_gemma.sh
+2
-2
No files found.
docs/start/install.rst
View file @
55093874
...
...
@@ -38,9 +38,8 @@ Image and tag: ``verlai/verl:vemlp-th2.4.0-cu124-vllm0.6.3-ray2.10-te1.7-v0.0.3`
.. code:: bash
# install the nightly version (recommended)
git clone https://github.com/volcengine/verl && cd verl && pip3 install -e .
# or install from pypi via `pip3 install verl`
# install the stable version
pip3 install verl
3. Setup Megatron (optional)
...
...
@@ -83,9 +82,7 @@ own post-training jobs.
# install verl together with some lightweight dependencies in setup.py
pip3 install torch==2.4.0 --index-url https://download.pytorch.org/whl/cu124
pip3 install flash-attn --no-build-isolation
git clone https://github.com/volcengine/verl.git
cd verl
pip3 install -e .
pip3 install verl
Megatron is optional. It's dependencies can be setup as below:
...
...
examples/ppo_trainer/run_gemma.sh
View file @
55093874
...
...
@@ -9,7 +9,7 @@ python3 -m verl.trainer.main_ppo \
data.max_response_length
=
512
\
actor_rollout_ref.model.path
=
google/gemma-2-2b-it
\
actor_rollout_ref.actor.optim.lr
=
1e-6
\
actor_rollout_ref.model.use_remove_padding
=
Tru
e
\
actor_rollout_ref.model.use_remove_padding
=
Fals
e
\
actor_rollout_ref.actor.ppo_mini_batch_size
=
128
\
actor_rollout_ref.actor.ppo_micro_batch_size_per_gpu
=
4
\
actor_rollout_ref.actor.fsdp_config.param_offload
=
False
\
...
...
@@ -22,7 +22,7 @@ python3 -m verl.trainer.main_ppo \
actor_rollout_ref.ref.log_prob_micro_batch_size_per_gpu
=
4
\
actor_rollout_ref.ref.fsdp_config.param_offload
=
True
\
critic.optim.lr
=
1e-5
\
critic.model.use_remove_padding
=
Tru
e
\
critic.model.use_remove_padding
=
Fals
e
\
critic.model.path
=
google/gemma-2-2b-it
\
critic.model.enable_gradient_checkpointing
=
False
\
critic.ppo_micro_batch_size_per_gpu
=
4
\
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment