Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
ZYM666 's Collections
Alignment

Alignment

updated Oct 13, 2024
Upvote
-

  • Direct Preference Optimization: Your Language Model is Secretly a Reward Model

    Paper • 2305.18290 • Published May 29, 2023 • 59

  • Towards Efficient and Exact Optimization of Language Model Alignment

    Paper • 2402.00856 • Published Feb 1, 2024

  • A General Theoretical Paradigm to Understand Learning from Human Preferences

    Paper • 2310.12036 • Published Oct 18, 2023 • 16

  • Statistical Rejection Sampling Improves Preference Optimization

    Paper • 2309.06657 • Published Sep 13, 2023 • 14

  • Direct Language Model Alignment from Online AI Feedback

    Paper • 2402.04792 • Published Feb 7, 2024 • 32
Upvote
-
  • Collection guide
  • Browse collections
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs