Skip to content

Conversation

@jacklanchantin
Copy link
Contributor

@jacklanchantin jacklanchantin commented Oct 14, 2025

What does this PR do? Please describe:

  • Adds GrpoLossConfig adv_std_normarlization (for DrGRPO)
  • Adds GrpoLossConfig loss_token_mean for normalizing over all tokens
  • Adds new if statement to skip ref logprob computation for kl if beta == 0
  • Adds tis_imp_ratio_cap to use truncated importance sampling correction

Fixes #{issue number}

Does your PR introduce any breaking changes? If yes, please list them:
List of all backwards-incompatible changes.

Check list:

  • Was the content of this PR discussed and approved via a GitHub issue? (no need for typos or documentation improvements)
  • Did you read the contributor guideline?
  • Did you make sure that your PR does only one thing instead of bundling different changes together?
  • Did you make sure to update the documentation with your changes? (if necessary)
  • Did you write any new necessary tests?
  • Did you verify new and existing tests pass locally with your changes?
  • Did you update the CHANGELOG? (no need for typos, documentation, or minor internal changes)

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Oct 14, 2025
@jacklanchantin jacklanchantin changed the title drgrpo Importance Sampling Correction, and DrGRPO args Oct 22, 2025
@jacklanchantin jacklanchantin changed the title Importance Sampling Correction, and DrGRPO args Change online training verifier Oct 22, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants