MedHallu: A Comprehensive Benchmark for Detecting Medical Hallucinations in Large Language Models Paper • 2502.14302 • Published 18 days ago • 9
CodeUpdateArena: Benchmarking Knowledge Editing on API Updates Paper • 2407.06249 • Published Jul 8, 2024
FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows" Paper • 2410.03727 • Published Sep 30, 2024 • 2
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression Paper • 2403.15447 • Published Mar 18, 2024 • 16
Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark Paper • 2402.11592 • Published Feb 18, 2024 • 2
Data-Free Knowledge Distillation for Heterogeneous Federated Learning Paper • 2105.10056 • Published May 20, 2021
DP-OPT: Make Large Language Model Your Privacy-Preserving Prompt Engineer Paper • 2312.03724 • Published Nov 27, 2023 • 1
Understanding Deep Gradient Leakage via Inversion Influence Functions Paper • 2309.13016 • Published Sep 22, 2023
Revisiting Data-Free Knowledge Distillation with Poisoned Teachers Paper • 2306.02368 • Published Jun 4, 2023