Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

AlignProp Support for direct reward finetuning #7312

Closed
parthos86 opened this issue Mar 14, 2024 · 3 comments
Closed

AlignProp Support for direct reward finetuning #7312

parthos86 opened this issue Mar 14, 2024 · 3 comments

Comments

@parthos86
Copy link

parthos86 commented Mar 14, 2024

Is your feature request related to a problem? Please describe.
No. AlignProp makes reward finetuning very fast compared to DDPO (about 25x) becoz of backpropagating the gradients directly from the reward function.

Describe the solution you'd like.
A similar integration to DDPO.
https://github.com/huggingface/diffusers/blob/main/docs/source/en/training/ddpo.md

Describe alternatives you've considered.
There is currently an implementation, however it is not well supported and is not within diffusers pipeline.
https://github.com/mihirp1998/AlignProp/

@mihirp1998

@parthos86 parthos86 changed the title AlignProp Support for direct reward finetuning https://github.com/mihirp1998/AlignProp/ AlignProp Support for direct reward finetuning Mar 14, 2024
@sayakpaul
Copy link
Member

It's more of trl thing really. Cc: @lvwerra @younesbelkada. diffusers is not a training focused library.

@mihirp1998
Copy link

I would be happy to do the the integration, although i would need some reference to what format/structure to follow

@sayakpaul
Copy link
Member

I am going to close this issue as it better belongs to the trl repo.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants