| # DeepCoder Experiment Runs |
|
|
| ## Command Snippets Overview |
|
|
| | Condition | Purpose | Variables | |
| |--------|---------|-----------| |
| | `GRPO top-p 1.0` | Full-retention baseline under linear top-p filtering | `rollout_filter_strategy=top_p`, `rollout_filter_value=1`, `rollout_filter_include_zero=True` | |
| | `GRPO top-p 0.9` | Stronger reward-variance filtering with adaptive retention | `rollout_filter_strategy=top_p`, `rollout_filter_value=0.9`, `rollout_filter_include_zero=False` | |
| | `GRPO top-k 0.25` | Fixed-budget filtering that keeps the top 25% of train groups | `rollout_filter_strategy=top_k`, `rollout_filter_value=0.25`, `rollout_filter_include_zero=True` | |
|
|
| All three command snippets run DeepCoder with `Qwen/Qwen2.5-Coder-7B`, `GRPO`, and single-turn code generation. |
|
|
| --- |
|
|
| ## 1. Top-p 1.0 (`Qwen/Qwen2.5-Coder-7B-200-GRPO-top-p-1`) |
|
|
| Uses linear top-p filtering with `rollout_filter_value=1`. |
|
|
| Goal: |
| - Establish a full-retention baseline while keeping the same reward-variance ranking machinery as the filtered runs |
|
|
| Key Details: |
| - Filtering uses `top_p`, `largest`, `reward_variance`, and `rollout_filter_top_p_prob_mode=linear` |
| - `rollout_filter_value=1` with `rollout_filter_include_zero=True` keeps the full train-group pool under linear top-p selection, so this is the closest thing to a no-filter baseline in `deepcoder_lines` |
| - `actor_rollout_ref.actor.use_ref=False` and `actor_rollout_ref.actor.use_kl_loss=False` remove reference-policy KL from training |
| - The run budget is `200` training steps with checkpoints every `20` steps |
|
|
| Source: |
| - `docs/deepcoder_lines`, lines `1-89` |
|
|
| Outputs: |
| - W&B project: `deepcoder_RAGEN_final_3` |
| - Run name: `Qwen/Qwen2.5-Coder-7B-200-GRPO-top-p-1` |
|
|
| --- |
|
|
| ## 2. Top-p 0.9 (`Qwen/Qwen2.5-Coder-7B-200-GRPO-top-p-0.9`) |
|
|
| Uses the same linear top-p filter, but keeps only the highest-variance groups whose score mass reaches `0.9`. |
|
|
| Goal: |
| - Increase reward-variance filtering strength while keeping the rest of the GRPO setup fixed |
|
|
| Key Details: |
| - Filtering again uses `top_p`, `largest`, `reward_variance`, and `rollout_filter_top_p_prob_mode=linear` |
| - `rollout_filter_value=0.9` makes retention adaptive: the number of kept groups depends on how reward variance is distributed across the `16` train groups |
| - `rollout_filter_include_zero=False` excludes zero-variance groups from selection |
| - Because `rollout_filter_type=largest`, the filter prioritizes groups with the highest within-group reward variance |
|
|
| Source: |
| - `docs/deepcoder_lines`, lines `93-181` |
|
|
| Outputs: |
| - W&B project: `deepcoder_RAGEN_final_3` |
| - Run name: `Qwen/Qwen2.5-Coder-7B-200-GRPO-top-p-0.9` |
|
|
| --- |
|
|
| ## 3. Top-k 0.25 (`Qwen/Qwen2.5-Coder-7B-200-GRPO-top-k-0.25`) |
|
|
| Switches from adaptive top-p filtering to fixed-fraction top-k filtering. |
|
|
| Goal: |
| - Compare adaptive top-p filtering against a fixed keep-top-25% regime |
|
|
| Key Details: |
| - `rollout_filter_strategy=top_k` with `rollout_filter_value=0.25` keeps `int(0.25 * 16) = 4` train groups per step |
| - With `es_manager.train.group_size=8`, this corresponds to at most `32` kept rollouts per training step after filtering |
| - `rollout_filter_include_zero=True` means zero-variance groups are still part of the ranking pool, but only the top `4` groups survive |
| - `rollout_filter_type=largest` means those `4` groups are chosen by highest reward variance |
|
|
| Source: |
| - `docs/deepcoder_lines`, lines `185-273` |
|
|
| Outputs: |
| - W&B project: `deepcoder_RAGEN_final_3` |
| - Run name: `Qwen/Qwen2.5-Coder-7B-200-GRPO-top-k-0.25` |
|
|
| --- |
|
|
| ## Common Notes |
|
|
| - Source format: |
| - `docs/deepcoder_lines` is a collection of three standalone bash snippets, not a parameterized sweep script |
| - The file defines both `USE_GRPO` and `USE_PPO`, but all three `python train.py` commands actually expand `$USE_GRPO` |
| - Shared setup across all three conditions: |
| - Config: `_10_deepcoder` |
| - Model: `Qwen/Qwen2.5-Coder-7B` |
| - `algorithm.adv_estimator=grpo` |
| - `agent_proxy.reward_normalization.method=identity` |
| - `trainer.total_training_steps=200` |
| - `ppo_mini_batch_size=32` |
| - `micro_batch_size_per_gpu=1` |
| - `es_manager.train.env_groups=16`, `es_manager.train.group_size=8` |
| - `es_manager.val.env_groups=256`, `es_manager.val.group_size=1` |
| - `trainer.n_gpus_per_node=8` |
| - `system.CUDA_VISIBLE_DEVICES="0,1,2,3,4,5,6,7"` |
| - `actor_rollout_ref.rollout.tensor_model_parallel_size=4` |
| - `agent_proxy.max_turn=1` |
| - `actor_rollout_ref.actor.use_ref=False` |
| - `actor_rollout_ref.rollout.rollout_filter_type=largest` |
| - `actor_rollout_ref.rollout.rollout_filter_metric=reward_variance` by default from `config/base.yaml` |
| - `actor_rollout_ref.rollout.rollout_filter_top_p_prob_mode=linear` |
| - `actor_rollout_ref.rollout.rollout_filter_empty_stop_steps=0` |
| - `actor_rollout_ref.rollout.max_model_len=10000` |
| - `actor_rollout_ref.rollout.max_num_batched_tokens=10000` |
| - `actor_rollout_ref.rollout.response_length=4000` |
| - `agent_proxy.fail_on_prompt_too_long=True` |
| - `lora.rank=0`, `lora.alpha=64`, `lora.target_modules=all-linear` |
| - `actor_rollout_ref.rollout.gpu_memory_utilization=0.6` |
| - `trainer.save_freq=20` |
| - `trainer.validation_steps=1` |
| - `trainer.val_before_train=True` |
| - `trainer.test_freq=10` |
| - `collapse_detection.first_turn_enabled=False` |
| - `collapse_detection.multi_turn_enabled=False` |
| - `trainer.resume_mode=disable` |
| - Logging and artifacts: |
| - Default local log dir remains `results/` |
| - Default logger remains `['console', 'wandb']` |
| - Checkpoints are saved every `20` steps |
|
|