ICLR 2025 "llm alignment" Papers
4 papers found
Is In-Context Learning Sufficient for Instruction Following in LLMs?
Hao Zhao, Maksym Andriushchenko, francesco croce et al.
ICLR 2025posterarXiv:2405.19874
21
citations
Learn Your Reference Model for Real Good Alignment
Alexey Gorbatovski, Boris Shaposhnikov, Alexey Malakhov et al.
ICLR 2025posterarXiv:2404.09656
46
citations
Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to Refuse
Maojia Song, Shang Hong Sim, Rishabh Bhardwaj et al.
ICLR 2025posterarXiv:2409.11242
29
citations
Uncovering Gaps in How Humans and LLMs Interpret Subjective Language
Erik Jones, Arjun Patrawala, Jacob Steinhardt
ICLR 2025posterarXiv:2503.04113
2
citations