nguyenhungquang / direct-preference-optimization Goto Github PK
View Code? Open in Web Editor NEWThis project forked from eric-mitchell/direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
License: Apache License 2.0