UnSeenTimeQA: Time-Sensitive Question-Answering Beyond LLMs' Memorization Paper • 2407.03525 • Published Jul 3, 2024 • 1
Triple Preference Optimization: Achieving Better Alignment with Less Data in a Single Step Optimization Paper • 2405.16681 • Published May 26, 2024 • 1
When "Competency" in Reasoning Opens the Door to Vulnerability: Jailbreaking LLMs via Novel Complex Ciphers Paper • 2402.10601 • Published Feb 16, 2024 • 1
How Can Input Reformulation Improve Tool Usage Accuracy in a Complex Dynamic Environment? A Study on $τ$-bench Paper • 2508.20931 • Published 10 days ago • 15
Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks Paper • 2404.14723 • Published Apr 23, 2024 • 10