#alignment 共 5 个条目 讲座 (2) L08: Post-training L16: AI's Impact on Humanity 论文 (3) AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback Direct Preference Optimization: Your Language Model is Secretly a Reward Model Training language models to follow instructions with human feedback