anthropics/hh-rlhf
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
MIT
Stargazers
- 8enmannSan Francisco
- AigizK
- alex-movilaFortech
- AlexPiche@mila-iqia
- amauboussinSan Francisco, CA
- brunoamaral@Lisbon-Collective
- daadaadaAnthropic
- DanielHesslowAdaptive
- danielmamayLondon
- denisfitz57
- donglixpMicrosoft Research
- esbenkcCopenhagen, Denmark
- fly51flyPRIS
- Fraser-GreenleeStealth
- freeman-lab@carbonplan
- gurusuraSura Systems Private Limited
- hkinkeMars
- KaixiangLinMICHIGAN STATE UNIVERSITY
- korymathGoogle DeepMind
- lxuechenStanford University
- mikestaubBright Machines
- ML-ChenStripe
- nawnoesSeoul, Korea
- orpheuslummis@CoincidenceNetwork
- pktsnffr
- pszemrajAtlanta, GA
- RobertKirk@ucl-dark
- rrmenon10UNC Chapel Hill
- samuelmcurtis@the-future-society
- saurav-anthropic
- stamate@Birkbeck-Computer-Science-Research
- stuhlmuellerElicit
- TheodoreGalanosAustrian Institute of Technology
- varunkumar-devAmazon
- vinuthams
- ZubinGouTsinghua University