Collections
Discover the best community collections!
Collections including paper arxiv:2309.15025
-
Large Language Model Alignment: A Survey
Paper • 2309.15025 • Published • 2 -
Aligning Large Language Models with Human: A Survey
Paper • 2307.12966 • Published • 1 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 48 -
SteerLM: Attribute Conditioned SFT as an (User-Steerable) Alternative to RLHF
Paper • 2310.05344 • Published • 1
-
Towards an Understanding of Large Language Models in Software Engineering Tasks
Paper • 2308.11396 • Published • 1 -
Several categories of Large Language Models (LLMs): A Short Survey
Paper • 2307.10188 • Published • 1 -
Large Language Models for Generative Recommendation: A Survey and Visionary Discussions
Paper • 2309.01157 • Published • 1 -
A Survey on Large Language Models for Recommendation
Paper • 2305.19860 • Published • 1
-
Moral Foundations of Large Language Models
Paper • 2310.15337 • Published • 1 -
Specific versus General Principles for Constitutional AI
Paper • 2310.13798 • Published • 2 -
Contrastive Prefence Learning: Learning from Human Feedback without RL
Paper • 2310.13639 • Published • 24 -
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 47