Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems Paper • 2502.19328 • Published 12 days ago • 21
ADELIE: Aligning Large Language Models on Information Extraction Paper • 2405.05008 • Published May 8, 2024 • 2
ADELIE Collection EMNLP2024 Main Conference: 《Aligning Large Language Models on Information Extraction》 • 7 items • Updated Nov 4, 2024 • 3