arXiv 2025
Authors: Hanjun Luo, Chiming Ni, Jiaheng Wen, et al., Hanan Salam.
Insight: We introduced a collaborative coding evaluation setup to quantify human-AI synergy end to end. Human-AI coding quality should be assessed through joint performance rather than isolated model generations.
NeurIPS 2025
Authors: Hanjun Luo, Shenyu Dai, Chiming Ni, et al., Hanan Salam.
Insight: We proposed AgentAuditor to test realistic agent behaviors across safety and security risk scenarios. Agent safety must be evaluated as an end-to-end system, not only by isolated model outputs.
EMNLP 2025 Main
Authors: Hanjun Luo, Yingbin Jin, Yiran Wang, et al., Zuozhu Liu.
Insight: We built DynamicNER with new data construction and evaluation protocols tailored to LLM-based NER. NER evaluation for LLMs benefits from dynamic, multilingual, and fine-grained settings beyond static benchmarks.
arXiv 2025
Authors: Kun Wang, Guibin Zhang, Zhenhong Zhou, et al., Hanjun Luo, others, Yang Liu.
Insight: We organized a full-stack taxonomy and synthesized mitigation strategies across the LLM/agent lifecycle. Safety risks propagate across the whole stack, so mitigation must align data, training, and deployment stages.
arXiv 2025
Authors: Hongyi Cai, Mohammad Mahdinur Rahman, Mingkang Dong, et al., Hanjun Luo, Yang Liu.
Insight: We implemented an automated framework that generates and applies debiasing interventions for text-to-image models. Automated debiasing pipelines can reduce social bias without full model retraining.