📚Papers

EMNLP 2025

Conference on Empirical Methods in Natural Language Processing

会议官网
915/ 3499 相关论文
Track
方向
Tier
915 / 915 篇论文
9
精读EMNLP 2025

Demystifying Synthetic Data in LLM Pre-training: A Systematic Study of Scaling Laws, Benefits, and Pitfalls

这篇论文要系统回答一个当前很现实的问题:合成数据放进 LLM 预训练里到底什么时候有用、收益怎么随规模变化、又会带来哪些坑。现有讨论对 synthetic pretraining data 的态度很分裂,有人把它当高质量数据缺口的补丁,有人担心会造成 model collapse 或分布退化;作者试图用 scaling law 式系统研究把这个问题讲清楚。

Feiyang Kang,Newsha Ardalani,Michael Kuchnik,Youssef Emad,Mostafa Elhoushi,Shubhabrata Sengupta,Shang-wen Li,Ramya Raghavendra,Ruoxi Jia,Carole-Jean Wu
synthetic-datascaling-lawdata-qualityDOIDBLP
9
精读EMNLP 2025

Not-Just-Scaling Laws: Towards a Better Understanding of the Downstream Impact of Language Model Design Decisions

这篇论文要解决的问题是:语言模型的下游表现并不由参数量和训练 token 数单独决定,但现有 scaling law 讨论通常把大量设计变量折叠掉了,导致我们很难判断哪些 pretrain 决策会稳定迁移到真实任务。作者试图把“模型设计选择→下游能力”的链条拆开,而不是继续把差异归因为单一规模因素。

Emmy Liu,Amanda Bertsch,Lintang Sutawika,Lindia Tjuatja,Patrick Fernandes,Lara Marinov ... 省略 2 位作者 ... ,Carolin Lawrence,Aditi Raghunathan,Kiril Gashteovski,Graham Neubig
scaling-lawdesign-choicesdownstreamDOIDBLP
9
精读EMNLP 2025

How to inject knowledge efficiently? Knowledge Infusion Scaling Law for Pre-training Large Language Models

Kangtao Lv,Haibin Chen,Yujin Yuan,Langming Liu,Shilei Liu,Yongwei Wang,Wenbo Su,Bo Zheng
pre-trainingscaling-lawknowledge-infusionDOIDBLP
9
精读EMNLP 2025

DSMoE: Matrix-Partitioned Experts with Dynamic Routing for Computation-Efficient Dense LLMs

Minxuan Lv,Zhenpeng Su,Leiyu Pan,Yizhe Xiong,Zijia Lin,Hui Chen ... 省略 3 位作者 ... ,Wenwu Ou,Di Zhang,Kun Gai,Songlin Hu
moearchitecturedynamic-routingDOIDBLP
9
精读FindingsEMNLP 2025

Instability in Downstream Task Performance During LLM Pretraining

Yuto Nishida,Masaru Isonuma,Yusuke Oda
pretrain-dynamicsdownstream-evalinstabilityAnthologyDBLP
9
精读EMNLP 2025

MIO: A Foundation Model on Multimodal Tokens

现有多模态大模型多采用外挂 encoder(如 CLIP)加 LLM 的架构,难以实现真正的跨模态原生统一建模。

Zekun Moore Wang,King Zhu,Chunpu Xu,Wangchunshu Zhou,Jiaheng Liu,Yibo Zhang ... 省略 7 位作者 ... ,Ge Zhang,Ke Xu,Jie Fu,Wenhao Huang
multimodal-tokensunified-modelfoundation-modelDOIDBLP
9
精读FindingsEMNLP 2025

SampleMix: A Sample-wise Pre-training Data Mixing Strategy by Coordinating Data Quality and Diversity

这篇论文要解决的是:预训练数据混合不能只看质量,也不能只看多样性,二者失衡都会伤害最终模型。以往配比策略常在 corpus 级别做固定权重,容易把高质量数据过采样到分布变窄,或把多样性拉满却引入过多噪声,因此 sample 级别的混合仍是一个悬而未决的问题。

Xiangyu Xi,Deyang Kong,Jian Yang,Jiawei Yang,Zhengyu Chen,Wei Wang,Jingang Wang,Xunliang Cai,Shikun Zhang,Wei Ye
data-mixingdata-qualitydata-diversityAnthologyDBLP
9
精读EMNLP 2025

AdamS: Momentum Itself Can Be A Normalizer for LLM Pretraining and Post-training

Adam/AdamW 在 LLM 预训练和后训练里是默认优化器,但它维护两个状态(一阶动量 m 和二阶动量 v),显存占用与参数量同量级;Lion 等无 v 的尝试在稳定性上不如 AdamW。作者问:能不能只保留 m,用 m 自身当作 normalizer,去掉 v 这一整块状态?

Huishuai Zhang,Bohan Wang,Luoxin Chen
optimizerpretrainingpost-trainingDOIDBLP
8
精读EMNLP 2025

Judging Quality Across Languages: A Multilingual Approach to Pretraining Data Filtering with Language Models

多语预训练缺高质量开源数据。主流 pipeline(CCNet、启发式规则)跨语言迁移性差、规则一旦改就要重跑;用 LLM 打分质量好但算力不现实。核心问题:怎么在大规模多语数据上既要 LLM 级别的质量判断,又要能跑得起。

Mehdi Ali,Manuel Brack,Max Lübbering,Elias Wendt,Abbas Goher Khan,Richard Rutmann ... 省略 9 位作者 ... ,Joachim Köhler,Patrick Schramowski,Michael Fromm,Kristian Kersting
Fraunhofer IAISTU Darmstadtdata-qualitymultilingualdata-filteringDOIarXivDBLP
8
精读EMNLP 2025

Cost-Optimal Grouped-Query Attention for Long-Context Modeling

Yingfa Chen,Yutong Wu,Chenyang Song,Zhen Leng Thai,Xingyu Shen,Xu Han,Zhiyuan Liu,Maosong Sun
gqalong-contextattentionDOIDBLP
8
精读EMNLP 2025

The Strawberry Problem: Emergence of Character-level Understanding in Tokenized Language Models

子词分词器(如 BPE)将字符打包成不透明的 token,导致 LLM 在处理拼写、字符计数(如经典的“strawberry 有几个 r”)等基础字符级任务时表现极差。

Adrian Cosma,Stefan Ruseti,Emilian Radoi,Mihai Dascalu
tokenizationemergencecharacter-levelDOIDBLP
8
精读FindingsEMNLP 2025

Error Typing for Smarter Rewards: Improving Process Reward Models with Error-Aware Hierarchical Supervision

这篇论文要解决的是:过程奖励模型(PRM)对“哪里错、错成什么类型”缺乏细粒度感知,导致奖励信号粗糙,难以稳定引导长链推理。很多现有 PRM 只给步骤好坏二元分或整体偏好,这会把不同错误机制混在一起,使模型学不到针对性的修正方向。

Pala Tej Deep,Panshul Sharma,Amir Zadeh,Chuan Li,Soujanya Poria
prmreward-modelrlhfAnthologyDBLP
8
精读FindingsEMNLP 2025

From Characters to Tokens: Dynamic Grouping with Hierarchical BPE

BPE 这种静态子词分词在多语言、代码、罕见词上分布不均,且固定粒度——要么碎成字符浪费算力,要么大 chunk 学不到形态。已有 byte-level / character-level 路线(ByT5、MambaByte、MegaByte、SpaceByte)效率或效果都有代价。

Rares Dolga,Lucas Maystre,Tudor Berariu,David Barber
SpotifyUCLtokenizerbpehierarchicalAnthologyDBLP
8
精读EMNLP 2025

ICL CIPHERS: Quantifying "Learning" in In-Context Learning via Substitution Ciphers

学界对上下文学习(ICL)的本质存在争议:LLM 究竟是在 prompt 中“学习”了新的输入输出映射,还是仅仅通过 prompt 激活了预训练权重中的语义先验?

Zhouxiang Fang,Aayush Mishra,Muhan Gao,Anqi Liu,Daniel Khashabi
Johns Hopkins Universityiclcipherslearning-analysisDOIDBLP
8
精读FindingsEMNLP 2025

MT-R1-Zero: Advancing LLM-based Machine Translation via R1-Zero-like Reinforcement Learning

这篇工作要解决的是:基于 LLM 的机器翻译虽然能通过 SFT 获得不错质量,但在忠实性、术语一致性和难例推理上仍受限,作者尝试把 R1-Zero 类强化学习范式迁移到 MT。过去 MT 上的 RL 多依赖句级奖励或人工设计指标,往往训练不稳、奖励窄、泛化有限。

Zhaopeng Feng,Shaosheng Cao,Jiahan Ren,Jiayuan Su,Ruizhe Chen,Yan Zhang,Jian Wu,Zuozhu Liu
r1-zerorltranslationAnthologyDBLP
8
精读FindingsEMNLP 2025

Nexus: Adaptive Upcycling to Efficiently Pretrain Mixture of Experts

Nikolas Gritsch,Qizhen Zhang,Acyr Locatelli,Sara Hooker,Ahmet Üstün
moeupcyclingpretrain-efficiencyAnthologyDBLP
8
精读EMNLP 2025

Reward-Weighted Sampling: Enhancing Non-Autoregressive Characteristics in Masked Diffusion LLMs

解决 Masked Diffusion LLMs (MDLM) 在非自回归(NAR)并行解码时,生成质量和连贯性不如传统自回归(AR)模型的问题。

Daehoon Gwak,Minseo Jung,Junwoo Park,Minho Park,ChaeHun Park,Junha Hyung,Jaegul Choo
diffusion-lmmasked-lmnon-autoregressiveDOIDBLP
8
精读EMNLP 2025

R-BPE: Improving BPE-Tokenizers with Token Reuse

解决传统 BPE 分词器词表冗余度高、对相似形态词汇泛化能力差的问题。

Nancy Hamdan,Osama Rakan Al Mraikhat,Fadi A. Zaraket
tokenizerbpepre-trainingDOIDBLP
8
精读FindingsEMNLP 2025

InfiMM-WebMath-40B: Advancing Multimodal Pre-Training for Enhanced Mathematical Reasoning

解决多模态大模型(VLM)在数学推理预训练阶段缺乏大规模、高质量图文交错数据的问题。

Xiaotian Han,Yiren Jian,Xuefeng Hu,Haogeng Liu,Yiqi Wang,Qihang Fan ... 省略 1 位作者 ... ,Huaibo Huang,Ran He,Zhenheng Yang,Quanzeng You
multimodal-pretrainingmath-reasoningvlmAnthologyDBLP
8
精读EMNLP 2025

Rewarding the Unlikely: Lifting GRPO Beyond Distribution Sharpening

这篇论文要解决的是:GRPO 是否只能做‘分布锐化’,也就是把高回报 token 序列进一步推高,而难以真正奖励低概率但更优的解。现有 group-based RL 常被批评为偏保守,只会在当前策略附近局部重排,导致探索不足和 mode collapse。

Andre Wang He,Daniel Fried,Sean Welleck
rlgrporeasoningDOIDBLP
8
精读EMNLP 2025

Pre-trained Models Perform the Best When Token Distributions Follow Zipf's Law

这篇工作的核心问题是:预训练语料的 token 分布是否存在一个比“越接近原始数据越好”更具体的最优形态,而作者给出的答案是 Zipf 分布附近最好。过去数据配比和采样通常凭经验做温度混合、去重和重加权,但很少把最终 token 频率形状本身当成一阶设计变量来系统分析。

Yanjin He,Qingkai Zeng,Meng Jiang
scaling-lawtoken-distributionzipfDOIDBLP
8
精读EMNLP 2025

Variance Sensitivity Induces Attention Entropy Collapse and Instability in Transformers

这篇工作要解释的是 transformer 训练中的一个核心稳定性问题:为什么某些设置下注意力会突然变得极低熵、极端尖锐,并进一步诱发训练不稳定。以往很多稳定性分析会把 loss spike 归因于学习率、归一化或数值精度,但对 attention entropy collapse 的直接成因往往缺少清晰机制。

Jonghyun Hong,Sungyoon Lee
attentiontraining-stabilityentropy-collapseDOIDBLP
8
精读FindingsEMNLP 2025

PEPE: Long-context Extension for Large Language Models via Periodic Extrapolation Positional Encodings

Jikun Hu,Dongsheng Guo,Yuli Liu,Qingyao Ai,Lixuan Wang,Xuebing Sun,Qilei Zhang,Quan Zhou,Cheng Luo
long-contextpositional-encodingextrapolationAnthologyDBLP
8
精读EMNLP 2025

OpenRLHF: A Ray-based Easy-to-use, Scalable and High-performance RLHF Framework

Jian Hu,Xibin Wu,Wei Shen,Jason Klein Liu,Weixun Wang,Songlin Jiang ... 省略 4 位作者 ... ,Xianyu,Yu Cao,Haotian Xu,Yiming Liu
rlhfframeworkscalableDOIDBLP
8
精读EMNLP 2025

Neuron-Level Differentiation of Memorization and Generalization in Large Language Models

Ko-Wei Huang,Yi-Fu Fu,Ching-Yu Tsai,Yu-Chieh Tu,Tzu-ling Cheng,Cheng-Yu Lin ... 省略 1 位作者 ... ,Heng-Yi Liu,Keng-Te Liao,Da-Cheng Juan,Shou-De Lin
memorizationgeneralizationinterpretabilityDOIDBLP
8
精读EMNLP 2025

Spectral Scaling Laws in Language Models: emphHow Effectively Do Feed-Forward Networks Use Their Latent Space?

探究随着模型规模扩大,前馈神经网络(FFN)是如何利用其高维隐空间的,是否存在可预测的规律。

Nandan Kumar Jha,Brandon Reagen
NYUscaling-lawspectral-analysisffnDOIDBLP
8
精读EMNLP 2025

TokenSmith: Streamlining Data Editing, Search, and Inspection for Large-Scale Language Model Training and Interpretability

Mohammad Aflah Khan,Ameya Godbole,Johnny Tian-Zheng Wei,Ryan Yixiang Wang,James Flemings,Krishna P. Gummadi,Willie Neiswanger,Robin Jia
pretrain-datatoolinginterpretabilityDOIDBLP
8
精读EMNLP 2025

Conditional [MASK] Discrete Diffusion Language Model

Hyukhun Koh,Minha Jhang,Dohyung Kim,Sangmook Lee,Kyomin Jung
diffusion-lmmasked-lmnon-arDOIDBLP
8
精读EMNLP 2025

Conan-Embedding-v2: Training an LLM from Scratch for Text Embeddings

现在主流 embedding 模型都是基于现成 LLM 做对比微调(BGE、E5、GTE 都是这条路),没人从头训一个 LLM 专门为 embedding 服务——这条路到底行不行、有没有额外增益,是个悬而未决的问题。

Shiyu Li,Yang Tang,Ruijie Liu,Shi-Zhe Chen,Xi Chen
Tencentembeddingspretrainingfrom-scratchDOIDBLP
8
精读EMNLP 2025

CoLA: Compute-Efficient Pre-Training of LLMs via Low-Rank Activation

Ziyue Liu,Ruijie Zhang,Zhengyang Wang,Mingsong Yan,Zi Yang,Paul D. Hovland,Bogdan Nicolae,Franck Cappello,Sui Tang,Zheng Zhang
low-rankactivationpretraining-efficiencyDOIDBLP
8
精读FindingsEMNLP 2025

Scaling Laws Are Unreliable for Downstream Tasks: A Reality Check

Nicholas Lourie,Michael Y. Hu,Kyunghyun Cho
scaling-lawevaluationdownstream-tasksAnthologyDBLP
8
精读EMNLP 2025

VersaTune: An Efficient Data Composition Framework for Training Multi-Capability LLMs

Keer Lu,Keshi Zhao,Zhuoran Zhang,Zheng Liang,Bin Cui,Tengjiao Wang,Wentao Zhang
data-mixturedata-compositionsftDOIDBLP
8
精读EMNLP 2025

Through the Valley: Path to Effective Long CoT Training for Small Language Models

Renjie Luo,Jiaxi Li,Chen Huang,Wei Lu
long-cotreasoningslmDOIDBLP
8
精读EMNLP 2025

s1: Simple test-time scaling

Niklas Muennighoff,Zitong Yang,Weijia Shi,Xiang Lisa Li,Li Fei-Fei,Hannaneh Hajishirzi,Luke Zettlemoyer,Percy Liang,Emmanuel J. Candès,Tatsunori Hashimoto
test-time-scalingreasoningsftDOIDBLP
8
精读EMNLP 2025

SWAN: An Efficient and Scalable Approach for Long-Context Language Modeling

标准 Transformer 的全局注意力在处理长上下文时面临二次方计算复杂度,导致长文本语言建模(Long-Context LM)效率低下且难以扩展。

Krishna C. Puvvada,Faisal Ladhak,Santiago Akle Serano,Cheng-Ping Hsieh,Shantanu Acharya,Somshubra Majumdar ... 省略 1 位作者 ... ,Samuel Kriman,Simeng Sun,Dima Rekesh,Boris Ginsburg
Applelong-contextlanguage-modelingscalingDOIDBLP
8
精读EMNLP 2025

DIDS: Domain Impact-aware Data Sampling for Large Language Model Training

预训练中的数据配比(Data Mixing)通常是静态的或基于简单的启发式规则,忽略了训练动态中不同领域数据之间复杂的正向与负向迁移(Negative Transfer)效应。

Weijie Shi,Jipeng Zhang,Yaguang Wu,Jingzhi Fang,Shibo Zhang,Yao Zhao ... 省略 3 位作者 ... ,Jia Zhu,Sirui Han,Jiajie Xu,Xiaofang Zhou
data-samplingdata-mixturepretrainingDOIDBLP
8
精读FindingsEMNLP 2025

FastCuRL: Curriculum Reinforcement Learning with Stage-wise Context Scaling for Efficient Training R1-like Reasoning Models

Mingyang Song,Mao Zheng,Zheng Li,Wenjie Yang,Xuan Luo
reinforcement-learningreasoningcurriculumAnthologyDBLP
8
精读IndustryEMNLP 2025

RLHF Algorithms Ranked: An Extensive Evaluation Across Diverse Tasks, Rewards, and Hyperparameters

Lucas Spangher,Rama Kumar Pasumarthi,Nick Masiewicki,William F. Arnold,Aditi Kaushal,Dale Johnson,Peter Grabowski,Eugene Ie
rlhfevaluationrewardDOIDBLP
8
精读FindingsEMNLP 2025

Uncovering Scaling Laws for Large Language Models via Inverse Problems

这篇论文要解决的是:现有 LLM scaling law 多是正向拟合,描述 loss 如何随参数、数据、算力变化,但难以从目标性能反推最优资源配置。工业上真正常见的问题是 inverse problem:给定 compute 或目标 loss,该怎么反算模型、数据和训练步数,而传统经验公式往往不稳定或只在局部区间有效。

Arun Verma,Zhaoxuan Wu,Zijian Zhou,Xiaoqiang Lin,Zhiliang Chen,Rachael Hwee Ling Sim ... 省略 8 位作者 ... ,Xinyi Xu,Apivich Hemachandra,See-Kiong Ng,Bryan Kian Hsiang Low
scaling-lawinverse-problemllm-capabilityAnthologyDBLP
8
精读EMNLP 2025

VocalNet: Speech LLMs with Multi-Token Prediction for Faster and High-Quality Generation

这篇论文要解决的是:Speech LLM 的自回归生成速度太慢,而简单并行化往往会损伤语音自然度和一致性。语音序列比文本长得多,逐 token 解码的时延尤其难接受,所以把文本 LLM 里的 multi-token prediction 思路迁到语音生成很有现实意义。

Yuhao Wang,Heyang Liu,Ziyang Cheng,Ronghua Wu,Qunshan Gu,Yanfeng Wang,Yu Wang
speech-lmmulti-token-predictiondecodingDOIDBLP
8
精读EMNLP 2025

Speech Discrete Tokens or Continuous Features? A Comparative Analysis for Spoken Language Understanding in SpeechLLMs

这篇论文要解决的是:SpeechLLM 做 spoken language understanding 时,输入到底该用离散 speech tokens 还是连续声学特征,没有形成稳定共识。很多工作沿用语音生成体系里的离散 token,因为它方便统一建模;但理解任务未必需要强离散化,连续特征可能保留更多细粒度韵律和说话风格信息。

Dingdong Wang,Junan Li,Mingyu Cui,Dongchao Yang,Xueyuan Chen,Helen M. Meng
speech-lmtokenizerdiscrete-tokenDOIDBLP
8
精读EMNLP 2025

Multilingual Language Model Pretraining using Machine-translated Data

低资源语言的预训练语料从哪来。原生高质量网页语料对许多语言都不够,作者直接问:能否用机器翻译把英文高质量数据搬到目标语言,用翻译数据做大规模多语言预训练。

Jiayi Wang,Yao Lu,Maurice Weber,Max Ryabinin,David Ifeoluwa Adelani,Yihong Chen,Raphael Tang,Pontus Stenetorp
Together AIUCLmultilingualpretrainingdata-mixtureDOIDBLP
8
精读EMNLP 2025

OpenS2S: Advancing Fully Open-Source End-to-End Empathetic Large Speech Language Model

Chen Wang,Tianyu Peng,Wen Yang,Yinan Bai,Guangfu Wang,Jun Lin ... 省略 1 位作者 ... ,Lingxiang Wu,Jinqiao Wang,Chengqing Zong,Jiajun Zhang
speech-lmspeech-to-speechopen-sourceDOIDBLP
8
精读EMNLP 2025

HMoE: Heterogeneous Mixture of Experts for Language Modeling

An Wang,Xingwu Sun,Ruobing Xie,Shuaipeng Li,Jiaqi Zhu,Zhen Yang ... 省略 2 位作者 ... ,Zhanhui Kang,Di Wang,Naoaki Okazaki,Cheng-Zhong Xu
moelanguage-modelingarchitectureDOIDBLP
8
精读FindingsEMNLP 2025

Rethinking Data Selection at Scale: Random Selection is Almost All You Need

这篇论文要解决的是:大规模数据选择是否真的需要复杂筛选器,还是随机选已经接近最优。过去几年很多工作在数据选择上投入大量工程,用质量模型、去重、聚类和课程学习来挑数据,但这些方法的真实增益在大规模下是否还能覆盖其成本,一直缺少足够直接的检验。

Tingyu Xia,Bowen Yu,Kai Dang,An Yang,Yuan Wu,Yuan Tian,Yi Chang,Junyang Lin
data-selectionpretrain-datascalingAnthologyDBLP
8
精读EMNLP 2025

Word Salad Chopper: Reasoning Models Waste A Ton Of Decoding Budget On Useless Repetitions, Self-Knowingly

推理模型(o1/R1 类 long-CoT)经常在 decoding 里写大段重复或自我绕圈的 '词汤',白白烧 token 预算却不提升答案质量。更有意思的是,模型自己其实知道自己在重复。作者要把这部分浪费切掉。

Wenya Xie,Shaochen Zhong,Hoang Anh Duy Le,Zhaozhuo Xu,Jianwen Xie,Zirui Liu
reasoning-modelsdecodingrepetitionDOIDBLP
8
精读EMNLP 2025

Temporal Scaling Law for Large Language Models

Chinchilla 等 scaling law 描述的是'训练结束后 final loss 随 N、D 的关系',没刻画训练过程中 loss 如何随 step 演化。实践里我们非常想在训练早期就外推出'这跑到底会收敛到什么 loss'。作者要给出一个描述 loss-vs-step 动态的时间维 scaling law。

Yizhe Xiong,Xiansheng Chen,Xin Ye,Hui Chen,Zijia Lin,Haoran Lian ... 省略 1 位作者 ... ,Wei Huang,Jianwei Niu,Jungong Han,Guiguang Ding
scaling-lawtemporal-scalingpretrainingDOIDBLP
8
精读EMNLP 2025

HS-STaR: Hierarchical Sampling for Self-Taught Reasoners via Difficulty Estimation and Budget Reallocation

Feng Xiong,Hongling Xu,Yifei Wang,Runxi Cheng,Yong Wang,Xiangxiang Chu
starreasoningrlDOIDBLP
8
精读EMNLP 2025

FIRE: Flexible Integration of Data Quality Ratings for Effective Pretraining

Liangyu Xu,Xuemiao Zhang,Feiyu Duan,Sirui Wang,Rongxiang Weng,Jingang Wang,Xunliang Cai
pretrainingdata-qualitydata-mixtureDOIDBLP
8
精读EMNLP 2025

Analyzing the Effects of Supervised Fine-Tuning on Model Knowledge from Token and Parameter Levels

这篇工作要解决的是:SFT 到底如何改变模型知识,目前通常只从下游表现看结果,缺少 token 级和参数级的细粒度分析。大家知道 SFT 会带来能力迁移和遗忘,但不知道知识是被覆盖、重加权,还是只是解码偏好被改写。

Junjie Ye,Yuming Yang,Yang Nan,Shuo Li,Qi Zhang,Tao Gui,Xuanjing Huang,Peng Wang,Zhongchao Shi,Jianping Fan
sftknowledgetraining-dynamicsDOIDBLP
8
精读EMNLP 2025

Data-Efficient Selection via Grammatical Complexity in Continual Pre-training of Domain-Specific LLMs

Yizhou Ying,Geng Zhang,Cui Danxin,Chengyu Du,Guanglei Yue,Sihang Jiang,Jiaqing Liang,Yifei Fu,Hailin Hu,Yanghua Xiao
continual-pretraindata-selectiondomainDOIDBLP
8
精读FindingsEMNLP 2025

Tool Zero: Training Tool-Augmented LLMs via Pure RL from Scratch

Yirong Zeng,Xiao Ding,Yutai Hou,Yuxian Wang,Li Du,Juyi Dai ... 省略 2 位作者 ... ,Dandan Tu,Weiwen Liu,Bing Qin,Ting Liu
rl-pretraintool-usepure-rlAnthologyDBLP
8
精读EMNLP 2025

AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time

Junyu Zhang,Runpei Dong,Han Wang,Xuying Ning,Haoran Geng,Peihao Li ... 省略 1 位作者 ... ,Yutong Bai,Jitendra Malik,Saurabh Gupta,Huan Zhang
test-time-computereasoningsystem-1-system-2DOIDBLP
8
精读EMNLP 2025

Flexible-length Text Infilling for Discrete Diffusion Models

离散 diffusion 语言模型做文本 infilling 时,通常要求提前指定要填多长。这是一个很强的人为约束——真实 infilling 场景(代码补全、文档改写)里长度本身就是未知的。作者要让 discrete diffusion 支持任意长度的中间段生成。

Andrew Zhang,Anushka Sivakumar,Chia-Wei Tang,Chris Thomas
Virginia Techdiffusion-lmtext-infillingnon-autoregressiveDOIDBLP
8
精读EMNLP 2025

Diffusion vs. Autoregressive Language Models: A Text Embedding Perspective

Siyue Zhang,Yilun Zhao,Liyuan Geng,Arman Cohan,Anh Tuan Luu,Chen Zhao
diffusion-lmembeddingar-vs-diffusionDOIDBLP
8
精读FindingsEMNLP 2025

More Tokens, Lower Precision: Towards the Optimal Token-Precision Trade-off in KV Cache Compression

KV cache 压缩中,token 数量和量化精度之间存在最优权衡点,但此前工作要么只做 token 剪枝、要么只做 KV 量化,缺乏对两者联合优化的系统分析。在固定显存预算下,应该保留更多低精度 token 还是更少高精度 token?

Jiebin Zhang,Dawei Zhu,Yifan Song,Wenhao Wu,Chuqiao Kuang,Xiaoguang Li,Lifeng Shang,Qun Liu,Sujian Li
kv-cachequantizationprecisionAnthologyDBLP
8
精读EMNLP 2025

Self-Adjust Softmax

这篇论文要解决的是:标准 softmax 在大词表建模里对概率尾部和梯度形状的处理可能并不理想,进而限制训练效率、校准或泛化。这个问题长期被接受为“输出层固定部件”,多数工作只在采样或 loss 上做文章,而不是直接改 softmax 本身。

Chuanyang Zheng,Yihang Gao,Guoxuan Chen,Han Shi,Jing Xiong,Xiaozhe Ren,Chao Huang,Zhenguo Li,Yu Li
softmaxarchitecturetraining-dynamicsDOIDBLP
7
泛读EMNLP 2025

Improving Cross Lingual Transfer by Pretraining with Active Forgetting

从标题判断,这篇论文关注的是:预训练阶段如何通过“主动遗忘”改善跨语言迁移,而不是一味累积所有语言知识。这个问题之所以值得做,是因为多语言预训练常见现象是高资源语言占据容量,导致跨语言迁移和低资源适配受干扰;但当前给定信息没有摘要,无法确认作者具体瞄准的是参数干扰、词表冲突、还是训练顺序问题。

Divyanshu Aggarwal,Ashutosh Sathe,Sunayana Sitaram
multilingualpretrainforgettingDOIDBLP
7
泛读FindingsEMNLP 2025

DM-Codec: Distilling Multimodal Representations for Speech Tokenization

这篇论文要解决的是:语音离散化如果只保留 acoustic token 或 semantic token,都会丢掉一部分对统一语音建模很关键的信息,尤其是上下文表征缺失会显著拉高困惑度并降低生成质量。过去的统一 speech tokenization 工作虽然开始结合声学和语义,但往往默认上下文信息可以由下游 LM 自己补;作者认为这一步不能再省。

Md Mubtasim Ahasan,Md Fahim,Tasnim Mohiuddin,Akmmahbubur Rahman,Aman Chadha,Tariq Iqbal,M. Ashraful Amin,Md Mofijul Islam,Amin Ahsan Ali
speech-tokenizerspeech-lmdistillationAnthologyarXivDBLP
7
泛读EMNLP 2025

A Systematic Analysis of Base Model Choice for Reward Modeling

这篇论文要解决的是:奖励模型训练时,底座模型的选择经常被当成默认项处理,但它实际上会显著影响 RM 质量和后续 RLHF 效果。过去大家更关注 preference 数据、loss 形式和训练配方,较少系统比较 base model;在底座模型快速增多的阶段,这个决策已经不能靠惯性。

Kian Ahrabian,Pegah Jandaghi,Negar Mokhberian,Sai Praneeth Karimireddy,Jay Pujara
reward-modelrlhfbase-modelDOIarXivDBLP
7
泛读EMNLP 2025

From Language to Cognition: How LLMs Outgrow the Human Language Network

这篇论文要解决的是:LLM 为什么会与人脑语言网络的神经活动相似,这种 brain alignment 在训练过程中到底跟哪类能力一起长。过去很多工作报告过 LLM 与脑信号的相关性,但很少系统拆开‘形式语言能力’和‘功能性语言能力’各自的作用,也缺少沿训练过程的动态视角。

Badr AlKhamissi,Greta Tuckute,Yingtian Tang,Taha Osama A Binhuraib,Antoine Bosselut,Martin Schrimpf
brain-alignmenttraining-dynamicsscalingDOIarXivDBLP
7
精读FindingsEMNLP 2025

From Tower to Spire: Adding the Speech Modality to a Translation-Specialist LLM

把语音模态加到已经训练好的文本 LLM 上,通常要么代价高(从头训 speech LLM),要么会损伤原有文本能力。作者想验证:能不能把离散化语音当作'一门额外的翻译语言',用很少的语音数据接到多语翻译 LLM 上。

Kshitij Ambilduke,Ben Peters,Sonal Sannigrahi,Anil Keshwani,Tsz Kin Lam,Bruno Martins,André F. T. Martins,Marcely Zanon Boito
Instituto de TelecomunicaçõesUnbabelInstituto Superior Técnicospeech-lmspeech-discretizationcontinual-pretrainAnthologyarXivDBLP
7
泛读EMNLP 2025

Group-Aware Reinforcement Learning for Output Diversity in Large Language Models

GRPO 之类的 group-relative RL 在提升准确率的同时会让模型 mode collapse,只重复少数高分回答。问题是 reward 只在 per-sample 层面算,group 内部的多样性没有被当成信号。

Oron Anschel,Alon Shoshan,Adam Botach,Shunit Haviv Hakimi,Asaf Gendler,Emanuel Ben Baruch,Nadav Bhonker,Igor Kviatkovsky,Manoj Aggarwal,Gérard G. Medioni
AmazonrlgrpodiversityDOIarXivDBLP
7
精读EMNLP 2025

A Good Plan is Hard to Find: Aligning Models with Preferences is Misaligned with What Helps Users

当前的对齐方法(如 RLHF)和评估基准(如 ChatbotArena)假设“用户偏好的回答”等同于“对用户完成任务有帮助的回答”,但在复杂规划任务中这一假设可能失效。

Nishant Balepur,Matthew Shu,Yoo Yeon Sung,Seraphina Goldfarb-Tarrant,Shi Feng,Fumeng Yang,Rachel Rudinger,Jordan Lee Boyd-Graber
rlhfalignment-taxreward-signalDOIarXivDBLP
7
精读EMNLP 2025

Sycophancy Mitigation Through Reinforcement Learning with Uncertainty-Aware Adaptive Reasoning Trajectories

当前的对齐训练范式会无意中促长模型的“阿谀奉承(Sycophancy)”现象——即使人类提供的信息是事实错误的,模型也会倾向于迎合和赞同。

Mohammad Beigi,Ying Shen,Parshin Shojaee,Qifan Wang,Zichao Wang,Chandan K. Reddy,Ming Jin,Lifu Huang
rlhfsycophancyreasoning-trajectoriesDOIarXivDBLP
7
精读EMNLP 2025

The Validation Gap: A Mechanistic Analysis of How Language Models Compute Arithmetic but Fail to Validate It

LLM 能够生成正确的算术步骤,但在被要求验证或自我纠错时却经常失败。这种“生成与验证能力脱节”的底层机制尚不清楚。

Leonardo Bertolazzi,Philipp Mondorf,Barbara Plank,Raffaella Bernardi
mechanistic-interpretabilityself-correctionarithmetic-reasoningDOIarXivDBLP
7
泛读EMNLP 2025

AlignX: Advancing Multilingual Large Language Models with Multilingual Representation Alignment

多语言 LLM 在不同语言间的表示空间存在错位,导致跨语言迁移能力弱、低资源语言表现差。

Mengyu Bu,Shaolei Zhang,Zhongjun He,Hua Wu,Yang Feng
BaiduICT CASmultilingualrepresentationalignmentDOIDBLP
7
泛读EMNLP 2025

LangVAE and LangSpace: Building and Probing for Language Model VAEs

这篇工作要解决的是:语言 VAE 一直被自回归解码器的 posterior collapse(潜变量被忽略)和缺少可解释潜空间探针这两个问题卡住,导致它长期被当成“做得出来但不好用”的语言建模分支。作者试图同时把“能训练起来”和“能分析 latent 学到了什么”这两件事补齐,因此不只是提一个 VAE 结构,也把 LangSpace 作为系统化 probing 框架一起提出。

Danilo S. Carvalho,Yingji Zhang,Harriet Unsworth,André Freitas
vaelatentlanguage-modelingDOIDBLP
7
泛读EMNLP 2025

Why Do Some Inputs Break Low-Bit LLM Quantization?

这篇工作要解决的是:低比特量化的 LLM 常出现输入相关的灾难性退化,同一个模型在大多数 prompt 上正常,但在少数输入上突然崩。过去很多分析把量化误差看成近似均匀或静态的权重误差,作者则直接追问:到底是什么样的输入会把低比特模型打坏,以及机制在哪里。

Ting-Yun Chang,Muru Zhang,Jesse Thomason,Robin Jia
quantizationlow-bitanalysisDOIDBLP
7
泛读FindingsEMNLP 2025

Hardware-Aware Parallel Prompt Decoding for Memory-Efficient Acceleration of LLM Inference

Hao Mark Chen,Wayne Luk,Yiu Ka Fai Cedric,Rui Li,Konstantin Mishchenko,Stylianos I. Venieris,Hongxiang Fan
speculative-decodingparallel-decodinginferenceAnthologyDBLP
7
泛读EMNLP 2025

Circuit Complexity Bounds for RoPE-based Transformer Architecture

Bo Chen,Xiaoyu Li,Yingyu Liang,Jiangxuan Long,Zhenmei Shi,Zhao Song,Jiahao Zhang
ropecircuit-complexitytheoryDOIDBLP
7
泛读EMNLP 2025

Faster In-Context Learning for LLMs via N-Gram Trie Speculative Decoding

Jinglin Chen,Qiwei Li,Zuchao Li,Baoyuan Qi,Guoming Liu,Haojun Ai,Hai Zhao,Ping Wang
speculative-decodingn-graminferenceDOIDBLP
7
泛读FindingsEMNLP 2025

Scale Down to Speed Up: Dynamic Data Selection for Reinforcement Learning

Zhuoyue Chen,Jihai Zhang,Ben Liu,Fangquan Lin,Wotao Yin
rldata-selectionrlvrAnthologyDBLP
7
泛读FindingsEMNLP 2025

On Domain-Adaptive Post-Training for Multimodal Large Language Models

Daixuan Cheng,Shaohan Huang,Ziyu Zhu,Xintong Zhang,Xin Zhao,Zhongzhi Luan,Bo Dai,Zhenliang Zhang
domain-adaptationcontinual-pretrainmllmAnthologyDBLP
7
泛读FindingsEMNLP 2025

Rethinking DPO: The Role of Rejected Responses in Preference Misalignment

DPO 中 rejected response 的角色被低估了——rejected 数据的质量和特性如何影响偏好对齐的效果,以及 DPO 中偏好错位(preference misalignment)的根源。

Jae Hyeon Cho,JunHyeok Oh,Myunsoo Kim,Byung-Jun Lee
dpopreference-alignmentrejected-responsesAnthologyDBLP
7
精读EMNLP 2025

Do All Autoregressive Transformers Remember Facts the Same Way? A Cross-Architecture Analysis of Recall Mechanisms

不同架构的自回归 Transformer(如 GPT、LLaMA 等变体)在事实回忆(factual recall)时是否使用相同的内部机制?此前的 mechanistic interpretability 研究多集中在单一架构上。

Minyeong Choe,Haehyun Cho,Changho Seo,Hyunil Kim
factual-recallcross-architectureinterpretabilityDOIDBLP
7
精读FindingsEMNLP 2025

Mamba Drafters for Speculative Decoding

在投机解码(Speculative Decoding)中,使用小 Transformer 作为草稿模型会面临 KV-cache 显存瓶颈,且随着上下文变长,草稿生成的加速比会急剧下降。

Daewon Choi,Seunghyuk Oh,Saket Dingliwal,Jihoon Tack,Kyuyoung Kim,Woomin Song ... 省略 2 位作者 ... ,Jinwoo Shin,Aram Galstyan,Shubham Katiyar,Sravan Babu Bodapati
speculative-decodingmambainference-efficiencyAnthologyDBLP
7
精读FindingsEMNLP 2025

Rotate, Clip, and Partition: Towards W2A4KV4 Quantization by Integrating Rotation and Learnable Non-uniform Quantizer

将 LLM 压缩到极低比特(如 W2A4KV4)时,激活值中的离群点(outliers)和权重分布的不均匀性会导致模型困惑度(Perplexity)崩溃。

Euntae Choi,Sumin Song,Woosang Lim,Sungjoo Yoo
quantizationrotationkv-cacheAnthologyDBLP
7
精读FindingsEMNLP 2025

SparsePO: Controlling Preference Alignment of LLMs via Sparse Token Masks

标准的偏好优化(如 DPO)对回复中的所有 token 均匀计算损失,但实际上决定人类偏好的往往只是少数几个关键 token(如某个事实错误或冒犯性词汇),全局计算会导致模型过度拟合噪声并遗忘通用能力。

Fenia Christopoulou,Ronald Cardenas,Gerasimos Lampouras,Haitham Bou-Ammar,Jun Wang
preference-optimizationsparse-tokenalignmentAnthologyDBLP
7
精读FindingsEMNLP 2025

Improving Influence-based Instruction Tuning Data Selection for Balanced Learning of Diverse Capabilities

这篇工作关注 instruction tuning 数据选择中的一个老问题:按 influence 选数据虽然能提升平均性能,但常常把模型推向少数强势能力,导致能力分布失衡。过去很多方法默认“高影响样本越多越好”,对多能力场景中的覆盖和冲突处理不够细;这篇论文显然是在修这个缺口。

Qirun Dai,Dylan Zhang,Jiaqi W. Ma,Hao Peng
data-selectioninstruction-tuninginfluence-functionsAnthologyDBLP
7
精读FindingsEMNLP 2025

Beyond Binary Preferences: Semi-Online Label-Free GRACE-KTO with Group-Wise Adaptive Calibration for High-Quality Long-Text Generation

这篇论文解决的是长文本生成偏好优化里的两个痛点:二元偏好信号太粗,且高质量标签昂贵,导致 KTO 这类方法在长文本上难以稳定学到细粒度质量差异。此前很多方法沿用 pairwise win/lose 标注,但长文本常常不是“全好或全坏”,而是局部段落优劣并存,因此二元标签天然丢信息。

Jingyang Deng,Ran Chen,Jo-Ku Cheng,Jinwen Ma
ktoalignmentlong-textAnthologyDBLP
7
精读FindingsEMNLP 2025

When Inverse Data Outperforms: Exploring the Pitfalls of Mixed Data in Multi-Stage Fine-Tuning

这篇论文关注 multi-stage fine-tuning 中一个常被低估的问题:把正向数据和所谓 inverse data 混在一起,不一定比只用 inverse data 更好。很多多阶段训练默认“数据越多、混合越稳”,但阶段间目标冲突、梯度抵消和分布污染可能让 mixed data 反而伤害最终效果。

Mengyi Deng,Xin Li,Tingyu Zhu,Zhicheng Yang,Zhijiang Guo,Wei Wang
data-mixturefine-tuningtraining-dynamicsAnthologyDBLP
7
精读FindingsEMNLP 2025

Identifying Rare Languages in Common Crawl Data is a Needles-in-a-Haystack Problem

Common Crawl 里低资源语言的识别是大海捞针:现有 LID(language ID)模型在稀有语言上误判率极高,导致预训练语料里这些语言要么被漏掉要么被噪声淹没。作者系统性量化了这个问题。

Rasul Dent,Pedro Ortiz Suarez,Thibault Clérice,Benoît Sagot
Inriadata-curationcommon-crawlmultilingualAnthologyDBLP
7
精读EMNLP 2025

Training compute-optimal transformer encoder models

Chinchilla 式 compute-optimal scaling law 是在 decoder-only 自回归模型上推的,encoder-only(BERT 类)的最优 data/param 配比是多少,一直没人系统做。作者把这个坑填了。

Megi Dervishi,Alexandre Allauzen,Gabriel Synnaeve,Yann LeCun
Meta FAIRscaling-lawcompute-optimalencoderDOIDBLP
7
精读IndustryEMNLP 2025

ECHO-LLaMA: Efficient Caching for High-Performance LLaMA Training

LLaMA 预训练的算力瓶颈很大一部分花在 attention 的重复计算上。作者想在训练期(不是推理期)引入某种 caching 机制,削减冗余计算。

Maryam Dialameh,Rezaul Karim,Hossein Rajabzadeh,Omar Mohamed Awad,Boxing Chen,Hyock Ju Kwon,Walid Ahmed,Yang Liu
training-efficiencycachingllamaDOIDBLP
7
精读EMNLP 2025

SoundMind: RL-Incentivized Logic Reasoning for Audio-Language Models

音频-语言模型在'需要逻辑推理'的音频任务上很弱:模型能听懂局部声学事件,但串不起因果/时序推理链。作者把 R1-style RL 迁到 ALM 上。

Xingjian Diao,Chunhui Zhang,Keyi Kong,Weiyi Wu,Chiyu Ma,Zhongyu Ouyang,Peijun Qing,Soroush Vosoughi,Jiang Gui
audio-lmrlreasoningDOIDBLP
7
精读EMNLP 2025

Machine-generated text detection prevents language model collapse

模型生成文本进入训练数据会引发 model collapse(Shumailov 等)。作者问:如果在数据管线前端加 MGT(machine-generated text)检测器,把合成数据过滤掉,是否能阻止 collapse。

George Drayson,Emine Yilmaz,Vasileios Lampos
UCLsynthetic-datamodel-collapsedata-qualityDOIDBLP
7
精读FindingsEMNLP 2025

Context Length Alone Hurts LLM Performance Despite Perfect Retrieval

Yufeng Du,Minyang Tian,Srikanth Ronanki,Subendhu Rongali,Sravan Babu Bodapati,Aram Galstyan,Azton Wells,Roy Schwartz,Eliu A. Huerta,Hao Peng
long-contextretrievalcontext-lengthAnthologyDBLP
7
精读FindingsEMNLP 2025

ConciseRL: Conciseness-Guided Reinforcement Learning for Efficient Reasoning Models

Razvan-Gabriel Dumitru,Darius Peteleaza,Vikas Yadav,Liangming Pan
reinforcement-learningreasoningefficiencyAnthologyDBLP
7
泛读FindingsEMNLP 2025

From Confidence to Collapse in LLM Factual Robustness

这篇工作要解决的是:LLM 的事实性并不是随置信度单调提升,模型可能先显得更自信,随后在分布外或扰动下突然崩塌。过去很多事实性评测只看平均准确率或校准误差,容易掩盖这种从 confidence 到 collapse 的非线性失稳。

Alina Fastowski,Bardh Prenkaj,Gjergji Kasneci
factualityrobustnessconfidenceAnthologyDBLP
7
泛读IndustryEMNLP 2025

A Proactive Reliability Metric for Detecting Failures in Language Model Training

这篇工作要解决的是:语言模型训练中的失败往往在 loss 爆炸、性能掉点或训练中断后才被发现,代价太高,作者提出一个前瞻性的可靠性指标来提前预警。现有监控通常依赖 loss、梯度范数或下游验证集,但这些信号往往太晚、太粗,难以在真正崩坏前触发动作。

Maryam Fatima
training-stabilityfailure-detectionreliabilityAnthologyDBLP
7
泛读EMNLP 2025

Convergence and Divergence of Language Models under Different Random Seeds

这篇工作要解决的是:仅改变随机种子时,语言模型会在多大程度上收敛到相似的表示和行为,又会在哪些维度发生分化。这个问题过去常被平均性能掩盖,因为很多论文默认不同 seed 只带来小噪声,但这对可复现性、机制解释和 checkpoint 选择都不够。

Finlay Fehlauer,Kyle Mahowald,Tiago Pimentel
training-dynamicsrandom-seedconvergenceDOIDBLP
7
泛读FindingsEMNLP 2025

Masked Diffusion Captioning for Visual Feature Learning

这篇工作要解决的是:图像表征学习通常依赖对比学习或掩码重建,而图像描述又多被当成下游任务,作者反过来用 masked diffusion captioning 来学视觉特征。这个问题值得关注,因为如果 caption generation 能作为更强的视觉预训练目标,它可能比纯重建更偏语义,又比对比学习保留更多生成结构。

Chao Feng,Zihao Wei,Andrew Owens
diffusion-lmmasked-lmvisual-representationAnthologyDBLP
7
泛读EMNLP 2025

Exploring morphology-aware tokenization: A case study on Spanish language modeling

Alba Táboas García,Piotr Przybyla,Leo Wanner
tokenizermorphologyspanishDOIDBLP
7
泛读EMNLP 2025

On Pruning State-Space LLMs

Tamer Ghattas,Michael Hassid,Roy Schwartz
pruningstate-space-modelssmDOIDBLP
7
泛读EMNLP 2025

Split-Merge: Scalable and Memory-Efficient Merging of Expert LLMs

Sruthi Gorantla,Aditya Rawal,Devamanyu Hazarika,Kaixiang Lin,Mingyi Hong,Mahdi Namazifar
model-mergingmoeexpert-mergingDOIDBLP
7
精读EMNLP 2025

Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality

这篇论文要回答的是:SFT 对齐质量到底受哪些因素主导,尤其是数据、层选择和训练细节分别起多大作用。过去很多对齐经验来自小规模经验主义或单因素实验,结论常常彼此冲突,因此需要系统性的大规模对照。

Yuto Harada,Yusuke Yamauchi,Yusuke Oda,Yohei Oseki,Yusuke Miyao,Yu Takagi
sftalignmentdata-qualityDOIDBLP
7
精读EMNLP 2025

Beyond Correctness: Confidence-Aware Reward Modeling for Enhancing Large Language Model Reasoning

这篇论文要解决的是:当前 reasoning 奖励模型过度关注‘答对没答对’,却忽略模型自己的置信度信号,导致奖励在边界样本上过于粗糙。只用 correctness 监督会把高置信错误、低置信正确和稳定正确混为一谈,这不利于推理能力和校准一起提升。

Qianxi He,Qingyu Ren,Shanzhe Lei,Xuhong Wang,Yingchun Wang
reward-modelreasoningconfidenceDOIDBLP
7
泛读FindingsEMNLP 2025

Mitigating Sequential Dependencies: A Survey of Algorithms and Systems for Generation-Refinement Frameworks in Autoregressive Models

Yunhai Hu,Zining Liu,Zhenyuan Dong,Tianfan Peng,Bradley McDanel,Sai Qian Zhang
speculative-decodingsurveyautoregressiveAnthologyDBLP
7
泛读FindingsEMNLP 2025

Explaining Length Bias in LLM-Based Preference Evaluations

Zhengyu Hu,Linxin Song,Jieyu Zhang,Zheyuan Xiao,Tianfu Wang,Zhengyu Chen,Nicholas Jing Yuan,Jianxun Lian,Kaize Ding,Hui Xiong
length-biaspreferencereward-modelAnthologyDBLP
7
泛读FindingsEMNLP 2025

Fast Quiet-STaR: Thinking Without Thought Tokens

Wei Huang,Yizhe Xiong,Xin Ye,Zhijie Deng,Hui Chen,Zijia Lin,Guiguang Ding
reasoningquiet-starlatent-reasoningAnthologyDBLP
7
泛读EMNLP 2025

Pierce the Mists, Greet the Sky: Decipher Knowledge Overshadowing via Knowledge Circuit Analysis

Haoming Huang,Yibo Yan,Jiahao Huo,Xin Zou,Xinfeng Li,Kun Wang,Xuming Hu
knowledge-circuitinterpretabilityknowledgeDOIDBLP
7
泛读EMNLP 2025

Koel-TTS: Enhancing LLM based Speech Generation with Preference Alignment and Classifier Free Guidance

Shehzeen Samarah Hussain,Paarth Neekhara,Xuesong Yang,Edresson Casanova,Subhankar Ghosh,Roy Fejgin,Mikyas T. Desta,Rafael Valle,Jason Li
speech-lmttspreference-alignmentDOIDBLP
7
精读FindingsEMNLP 2025

How a Bilingual LM Becomes Bilingual: Tracing Internal Representations with Sparse Autoencoders

Tatsuro Inaba,Go Kamoda,Kentaro Inui,Masaru Isonuma,Yusuke Miyao,Yohei Oseki,Yu Takagi,Benjamin Heinzerling
bilingualsparse-autoencoderinterpretabilityAnthologyDBLP
7
精读EMNLP 2025

Improbable Bigrams Expose Vulnerabilities of Incomplete Tokens in Byte-Level Tokenizers

揭示字节级BPE分词器在处理不完整Token或极低频双字节组合时的结构性脆弱问题。

Eugene Jang,Kimin Lee,Jin-Woo Chung,Keuntae Park,Seungwon Shin
KAISTtokenizerbyte-levelrobustnessDOIDBLP
7
泛读FindingsEMNLP 2025

ZEBRA: Leveraging Model-Behavioral Knowledge for Zero-Annotation Preference Dataset Construction

这篇工作要解决的是:在没有人工偏好标注的情况下,如何构造可用于偏好学习的数据集。现有做法通常依赖人工对比标注或强教师模型打分,成本高且容易把教师偏好直接灌进数据;作者尝试改用“模型行为知识”来替代人工监督,但从给定信息里还看不到其行为信号如何定义、覆盖哪些偏好维度。

Jeesu Jung,Chanjun Park,Sangkeun Jung
preference-dataalignmentsynthetic-dataAnthologyDBLP
7
泛读EMNLP 2025

ToDi: Token-wise Distillation via Fine-Grained Divergence Control

这篇工作要解决的是:知识蒸馏里按序列整体对齐分布太粗,无法稳定地把教师模型的 token 级行为传给学生。传统 distillation 往往只在句级 logits 或最终目标上做匹配,结果是学生在难 token、长尾 token 和不确定位置上学得不均衡;作者提出 token-wise 的细粒度散度控制来修正这个问题。

Seongryong Jung,Suwan Yoon,DongGeon Kim,Hwanhee Lee
distillationtraining-objectivetoken-levelDOIDBLP
7
泛读EMNLP 2025

Code Execution as Grounded Supervision for LLM Reasoning

这篇工作聚焦一个很具体但重要的问题:LLM 的推理监督大多来自最终答案或人工过程标注,缺少可验证、低噪声的中间监督。尤其在数学和程序化推理里,仅靠 outcome supervision 容易让模型学会表面格式而不是正确过程;作者尝试把代码执行结果当作 grounded supervision 来约束推理。

Dongwon Jung,Wenxuan Zhou,Muhao Chen
code-executionreasoningsupervisionDOIDBLP
7
泛读FindingsEMNLP 2025

False Friends Are Not Foes: Investigating Vocabulary Overlap in Multilingual Language Models

这篇论文要解决的是:多语种模型里的词表重叠,尤其是“假朋友词”(拼写相近但语义不同),到底是在帮模型共享知识,还是在制造跨语言干扰。过去关于共享词表的讨论常把 overlap 视为天然有利,或简单担心 negative transfer;作者的结论从标题就很明确:false friends 并不一定是坏事。

Julie Kallini,Dan Jurafsky,Christopher Potts,Martijn Bartelds
multilingualtokenizervocabularyAnthologyDBLP
7
泛读EMNLP 2025

Investigating How Pre-training Data Leakage Affects Models' Reproduction and Detection Capabilities

这篇工作研究的是:预训练数据泄漏如何同时影响模型的“复现能力”和“检测能力”。过去关于 data leakage 的讨论多集中在 benchmark 污染或 memorization 本身,很少把“模型会不会复现泄漏内容”和“我们能不能检测到泄漏痕迹”放到一起系统分析;作者显然想把两者联系起来。

Masahiro Kaneko,Timothy Baldwin
data-leakagepretraining-dataevaluationDOIDBLP
7
泛读EMNLP 2025

ModRWKV: Transformer Multimodality in Linear Time

Jiale Kang,Ziyin Yue,Qingyu Yin,Rui Jiang,Weile Li,Zening Lu,Zhouran Ji
rwkvmultimodalarchitectureDOIDBLP
7
泛读EMNLP 2025

Tokenization and Representation Biases in Multilingual Models on Dialectal NLP Tasks

Vani Kanjirangat,Tanja Samardzic,Ljiljana Dolamic,Fabio Rinaldi
tokenizermultilingualbiasDOIDBLP
7
泛读EMNLP 2025

Multilingual Pretraining for Pixel Language Models

Ilker Kesen,Jonas F. Lotz,Ingo Ziegler,Phillip Rust,Desmond Elliott
pixel-lmmultilingualtokenizerDOIDBLP
7
泛读IndustryEMNLP 2025

LP Data Pipeline: Lightweight, Purpose-driven Data Pipeline for Large Language Models

Yungi Kim,Hyunsoo Ha,Seonghoon Yang,Sukyung Lee,Jihoo Kim,Chanjun Park
pretrain-datadata-pipelineDOIDBLP
7
泛读FindingsEMNLP 2025

FractalLLM: Lossless Self-Speculative Decoding with Layer Embedded Self-Compression

Juhyeong Kim,Sangyeon Yu,Gyunyeop Kim,Sangwoo Kang
speculative-decodingself-compressioninferenceAnthologyDBLP
7
泛读EMNLP 2025

Retracing the Past: LLMs Emit Training Data When They Get Lost

Myeongseob Ko,Nikhil Reddy Billa,Adam Nguyen,Charles Fleming,Ming Jin,Ruoxi Jia
memorizationtraining-dataanalysisDOIDBLP
7
泛读EMNLP 2025

Integral Transformer: Denoising Attention, Not Too Much Not Too Little

Ivan Kobyzev,Abbas Ghaddar,Dingtao Hu,Boxing Chen
attentionarchitectureDOIDBLP
7
泛读EMNLP 2025

Bitune: Leveraging Bidirectional Attention to Improve Decoder-Only LLMs

Dawid Jan Kopiczko,Tijmen Blankevoort,Yuki M. Asano
bidirectional-attentionarchitecturedecoderDOIDBLP
7
泛读EMNLP 2025

Recursive Training Loops in LLMs: How training data properties modulate distribution shift in generated data?

Grgur Kovac,Jérémy Perez,Rémy Portelas,Peter Ford Dominey,Pierre-Yves Oudeyer
synthetic-datatraining-dynamicsdistribution-shiftDOIDBLP
7
泛读EMNLP 2025

Data and Model Centric Approaches for Expansion of Large Language Models to New languages

Anoop Kunchukuttan,Raj Dabre,V. Rudra Murthy,Mohammed Safi Ur Rahman Khan,Thanmay Jayakumar
multilingualcontinual-pretrainlanguage-expansionDOIDBLP
7
精读FindingsEMNLP 2025

S*: Test Time Scaling for Code Generation

这篇论文要解决的是代码生成在 test-time 如何继续靠额外计算换正确率,而不只是靠更大的 base model。过去这类提升常依赖 pass@k 采样、执行反馈或搜索,但缺少一个清晰的“test-time scaling”框架来系统利用推理时预算。

Dacheng Li,Shiyi Cao,Chengkun Cao,Xiuyu Li,Shangyin Tan,Kurt Keutzer,Jiarong Xing,Joseph E. Gonzalez,Ion Stoica
test-time-scalingcode-generationinference-scalingAnthologyDBLP
7
精读EMNLP 2025

Search-o1: Agentic Search-Enhanced Large Reasoning Models

这篇论文要解决的是大推理模型在需要外部信息时的闭卷局限。纯靠内部参数做 long reasoning,遇到知识缺失、时效性问题或多跳事实链时,模型即使会推也可能无料可推。

Xiaoxi Li,Guanting Dong,Jiajie Jin,Yuyao Zhang,Yujia Zhou,Yutao Zhu,Peitian Zhang,Zhicheng Dou
reasoning-modelssearch-augmentationagentDOIDBLP
7
精读EMNLP 2025

APLOT: Robust Reward Modeling via Adaptive Preference Learning with Optimal Transport

这篇论文要解决的是 reward model 在偏好学习中的鲁棒性问题。标准偏好建模通常把成对偏好当作干净监督,但真实偏好数据里存在噪声、分布偏移和标注者异质性,导致 reward model 很容易学到不稳定甚至可被利用的偏好边界。

Zhuo Li,Yuege Feng,Dandan Guo,Jinpeng Hu,Anningzhe Gao,Xiang Wan
reward-modelingpreference-learningoptimal-transportDOIDBLP
7
精读EMNLP 2025

RED: Unleashing Token-Level Rewards from Holistic Feedback via Reward Redistribution

Jiahui Li,Lin Li,Tai-Wei Chang,Kun Kuang,Long Chen,Jun Zhou,Cheng Yang
rlhftoken-level-rewardreward-redistributionDOIDBLP
7
精读FindingsEMNLP 2025

FG-PRM: Fine-grained Hallucination Detection and Mitigation in Language Model Mathematical Reasoning

Ruosen Li,Ziming Luo,Xinya Du
prmmath-reasoninghallucinationAnthologyDBLP
7
泛读EMNLP 2025

Expectation Preference Optimization: Reliable Preference Estimation for Improving the Reasoning Capability of Large Language Models

DPO 类偏好优化在推理任务上经常翻车:同一个 prompt 下不同采样的 reward 方差极大,用单点奖励去估计偏好会引入很强的噪声。

Zelin Li,Dawei Song
preference-optimizationreasoningalignmentDOIDBLP
7
泛读EMNLP 2025

CoMMIT: Coordinated Multimodal Instruction Tuning

多模态指令微调里,视觉 encoder 和 LLM 的学习速度/梯度尺度不匹配,很多时候 LLM 已经过拟合指令格式,视觉侧还没动起来(或反过来),最终融合不充分。

Xintong Li,Junda Wu,Tong Yu,Rui Wang,Yu Wang,Xiang Chen,Jiuxiang Gu,Lina Yao,Julian J. McAuley,Jingbo Shang
multimodalinstruction-tuningcoordinationDOIDBLP
7
泛读EMNLP 2025

Dynamic Expert Specialization: Towards Catastrophic Forgetting-Free Multi-Domain MoE Adaptation

Junzhuo Li,Bo Wang,Xiuze Zhou,Xuming Hu
moecontinual-learningadaptationDOIDBLP
7
泛读FindingsEMNLP 2025

Do Influence Functions Work on Large Language Models?

影响函数(Influence Functions, IF)在理论上能精准归因训练数据对模型预测的贡献,但在参数量巨大、损失地形高度非凸的现代 LLM 上,其有效性和扩展性存疑。

Zhe Li,Wei Zhao,Yige Li,Jun Sun
influence-functionsdata-attributionAnthologyDBLP
7
泛读FindingsEMNLP 2025

Conv-Basis: A New Paradigm for Efficient Attention Inference and Gradient Computation in Transformers

标准 Transformer 的精确注意力计算具有 O(N^2) 的时间和空间复杂度,而现有的线性注意力近似方法往往在长序列上损失精度,且在训练时的梯度计算效率低下。

Yingyu Liang,Heshan Liu,Zhenmei Shi,Zhao Song,Zhuoyan Xu,Jiale Zhao,Zhen Zhuang
attentionefficient-attentiongradientAnthologyDBLP
7
泛读EMNLP 2025

Enhancing Efficiency and Exploration in Reinforcement Learning for LLMs

大模型在 RLHF(如 PPO)阶段常面临样本效率低下、方差大以及探索不足(容易陷入局部最优或模式崩溃)的问题。

Mengqi Liao,Xiangyu Xi,Ruinian Chen,Jia Leng,Yangen Hu,Ke Zeng,Shuai Liu,Huaiyu Wan
rlexplorationrlvrDOIDBLP
7
泛读EMNLP 2025

Efficient Model Development through Fine-tuning Transfer

每次面对新任务都从头微调大规模 LLM 成本极高,而小模型上积累的微调权重通常无法直接复用到大模型上。

Pin-Jie Lin,Rishab Balasubramanian,Fengyuan Liu,Nikhil Kandpal,Tu Vu
finetuning-transfermodel-developmentefficiencyDOIDBLP
7
泛读FindingsEMNLP 2025

Training with Fewer Bits: Unlocking Edge LLMs Training with Stochastic Rounding

在边缘设备上训练 LLM 受限于极低的显存,而传统的低比特量化(如 INT4)在训练时采用最近邻舍入,会破坏梯度的期望值,导致模型无法收敛。

Taowen Liu,Marta Andronic,Deniz Gündüz,George Anthony Constantinides
low-bitstochastic-roundingtrainingAnthologyDBLP
7
泛读EMNLP 2025

On Relation-Specific Neurons in Large Language Models

这篇论文研究的核心问题是:LLM 里是否存在与特定关系类型稳定对应的神经元,以及这些神经元到底是可解释机制还是只是相关性伪影。过去神经元解释工作常停留在词性、主题或单一事实激活层面,但“关系”是更接近知识组织单元的粒度。

Yihong Liu,Runsheng Chen,Lea Hirlimann,Ahmad Dawar Hakimi,Mingyang Wang,Amir Hossein Kargaran,Sascha Rothe,François Yvon,Hinrich Schütze
interpretabilityneuronsknowledgeDOIDBLP
7
泛读FindingsEMNLP 2025

LAWCAT: Efficient Distillation from Quadratic to Linear Attention with Convolution across Tokens for Long Context Modeling

这篇论文要解决的是长上下文建模里二次复杂度注意力成本过高,但很多线性注意力替代方案在蒸馏后质量掉得明显。作者试图把“从 full attention 蒸馏到 linear attention”这件事做得更有效,缩小效率和性能之间的缺口。

Zeyu Liu,Souvik Kundu,Lianghao Jiang,Anni Li,Srikanth Ronanki,Sravan Babu Bodapati,Gourav Datta,Peter Anthony Beerel
linear-attentiondistillationlong-contextAnthologyDBLP
7
泛读IndustryEMNLP 2025

Structural Reward Model: Enhancing Interpretability, Efficiency, and Scalability in Reward Modeling

这篇论文解决的是奖励模型在可解释性、训练成本和扩展性上的三难:标准 reward model 往往是黑盒打分器,数据一多和任务一变就难维护,也不容易知道它到底在奖励什么。作者提出 Structural Reward Model,试图用更结构化的方式表达奖励。

Xiaoyu Liu,Di Liang,Hongyu Shan,Peiyang Liu,Yonghao Liu,Muling Wu ... 省略 1 位作者 ... ,Xianjie Wu,Li Miao,Jiangrong Shen,Minlong Peng
reward-modelinterpretabilityDOIDBLP
7
泛读EMNLP 2025

NOVER: Incentive Training for Language Models via Verifier-Free Reinforcement Learning

Wei Liu,Siya Qi,Xinyu Wang,Chen Qian,Yali Du,Yulan He
rlverifier-freereward-modelDOIDBLP
7
精读FindingsEMNLP 2025

Tracing Multilingual Factual Knowledge Acquisition in Pretraining

Yihong Liu,Mingyang Wang,Amir Hossein Kargaran,Felicia Körner,Ercong Nie,Barbara Plank,François Yvon,Hinrich Schütze
multilingualfactual-knowledgepretraining-dynamicsAnthologyDBLP
7
精读EMNLP 2025

Think in Safety: Unveiling and Mitigating Safety Alignment Collapse in Multimodal Large Reasoning Model

Xinyue Lou,You Li,Jinan Xu,Xiangyu Shi,Chi Chen,Kaiyu Huang
multimodalreasoning-modelsafety-alignmentDOIDBLP
7
精读FindingsEMNLP 2025

RAISE: Reinforced Adaptive Instruction Selection For Large Language Models

Qingsong Lv,Yangning Li,Zihua Lan,Zishan Xu,Jiwei Tang,Tingwei Lu ... 省略 1 位作者 ... ,Wenhao Jiang,Hong-Gee Kim,Hai-Tao Zheng,Philip S. Yu
instruction-selectionrldata-qualityAnthologyDBLP
7
泛读FindingsEMNLP 2025

LegoSLM: Connecting LLM with Speech Encoder using CTC Posteriors

Rao Ma,Tongzhou Chen,Kartik Audhkhasi,Bhuvana Ramabhadran
speech-lmctcmultimodalAnthologyDBLP
7
泛读EMNLP 2025

UnitCoder: Scalable Code Synthesis from Pre-training Corpora

预训练阶段的高质量代码数据稀缺,现有方法高度依赖昂贵的大模型蒸馏(如 GPT-4 生成),难以低成本扩展到海量规模。

Yichuan Ma,Yunfan Shao,Peiji Li,Demin Song,Qipeng Guo,Linyang Li,Xipeng Qiu,Kai Chen
Shanghai AI LabFudan Universitycode-generationpretraining-datasynthesisDOIDBLP
7
泛读FindingsEMNLP 2025

Improving Preference Alignment of LLM with Inference-Free Self-Refinement

传统的自我反思(Self-Refinement)或偏好对齐方法在训练时需要模型在线自回归生成回复,推理开销极大,严重拖慢了 Post-train 的迭代节奏。

Fukun Ma,Kaibin Tian,Jieting Xue,Xiaoyi Wang,Ye Ma,Quan Chen,Peng Jiang,Lijie Wen
alignmentself-refinementpreference-optimizationAnthologyDBLP
7
泛读EMNLP 2025

Does quantization affect models' performance on long-context tasks?

这篇论文问的是一个很实际但常被忽略的问题:量化是否会以不同于短上下文基准的方式伤害长上下文能力。过去很多量化评测集中在常规困惑度或短任务准确率上,但长上下文检索、跨段整合和位置稳定性可能对数值误差更敏感。

Anmol Mekala,Anirudh Atmakuru,Yixiao Song,Marzena Karpinska,Mohit Iyyer
quantizationlong-contextevaluationDOIDBLP
7
泛读FindingsEMNLP 2025

Length Representations in Large Language Models

Sangjun Moon,Dasom Choi,Jingun Kwon,Hidetaka Kamigaito,Manabu Okumura
interpretabilitylengthrepresentationAnthologyDBLP
7
泛读EMNLP 2025

We Need to Measure Data Diversity in NLP - Better and Broader

Dong Nguyen,Esther Ploeger
data-diversitydata-qualityevaluationDOIDBLP
7
精读FindingsEMNLP 2025

Probing Semantic Routing in Large Mixture-of-Expert Models

Matthew Lyle Olson,Neale Ratzlaff,Musashi Hinck,Man Luo,Sungduk Yu,Chendi Xue,Vasudev Lal
moesemantic-routinginterpretabilityAnthologyDBLP
7
泛读EMNLP 2025

How do autoregressive transformers solve full addition?

Wang Peixu,Chen Yu,Yu Ming,Cheng Xiang
mechanistic-interpretabilityarithmetic-reasoningautoregressive-transformersDOIDBLP
7
泛读EMNLP 2025

Understanding Subword Compositionality of Large Language Models

Qiwei Peng,Yekun Chai,Anders Søgaard
tokenizersubword-compositionalitymechanistic-interpretabilityDOIDBLP
7
泛读EMNLP 2025

VerIF: Verification Engineering for Reinforcement Learning in Instruction Following

Hao Peng,Yunjia Qi,Xiaozhi Wang,Bin Xu,Lei Hou,Juanzi Li
rlhfinstruction-followingverificationDOIDBLP
7
精读FindingsEMNLP 2025

Bag of Tricks for Sparse Mixture-of-Experts: A Benchmark Across Reasoning, Efficiency, and Safety

这篇工作要解决的是 sparse MoE 领域缺少一套同时覆盖推理能力、效率和安全性的横向评测。过去很多 MoE 论文只强调训练 FLOPs、平均 benchmark 或路由稀疏度,但 MoE 真正部署时还会遇到 latency 抖动、长尾专家负载、对抗安全和 reasoning 稳定性等问题,单指标结论很容易误导。

Mufan Qiu,Zheyu Shen,Pingzhi Li,Ang Li,Tianlong Chen
moebenchmarkefficiencyAnthologyDBLP
7
泛读EMNLP 2025

Exploring the Limitations of Mamba in COPY and CoT Reasoning

Ruifeng Ren,Zhicong Li,Yong Liu
mambassmarchitectureDOIDBLP
7
泛读EMNLP 2025

Train It and Forget It: Merge Lists are Unnecessary for BPE Inference in Language Models

Tomohiro Sawada,Kartik Goyal
tokenizerbpeinferenceDOIDBLP
7
泛读FindingsEMNLP 2025

Assessing the Role of Data Quality in Training Bilingual Language Models

Skyler Seto,Maartje ter Hoeve,Maureen de Seyssel,David Grangier
data-qualitybilingualpretrainingAnthologyDBLP
7
泛读EMNLP 2025

R-PRM: Reasoning-Driven Process Reward Modeling

Shuaijie She,Junxiao Liu,Yifeng Liu,Jiajun Chen,Xin Huang,Shujian Huang
process-reward-modelreasoningrlDOIDBLP
7
精读FindingsEMNLP 2025

Linear Steerability in Language Models: When It Emerges and How It Evolves

Jianshu She,Xinyue Li,Eric P. Xing,Zhengzhong Liu,Qirong Ho
steerabilityrepresentationtraining-dynamicsAnthologyDBLP
7
泛读EMNLP 2025

From Unaligned to Aligned: Scaling Multilingual LLMs with Multi-Way Parallel Corpora

多语言 LLM 的核心瓶颈之一是低资源语言与高资源语言之间的能力差距。以往做法要么依赖机器翻译扩充平行语料,要么在 pretrain 阶段简单混合多语言数据,但缺乏系统性地利用多路平行语料(multi-way parallel corpora)来对齐多语言表示。

Yingli Shen,Wen Lai,Shuo Wang,Ge Gao,Kangyang Luo,Alexander Fraser,Maosong Sun
multilingualparallel-corporapretrainingDOIDBLP
7
精读EMNLP 2025

CODI: Compressing Chain-of-Thought into Continuous Space via Self-Distillation

Chain-of-Thought (CoT) 推理虽然提升了 LLM 的推理能力,但显式生成长推理链带来巨大的推理延迟和 token 开销。本文要解决的是:能否把 CoT 推理过程压缩到连续隐空间中,在不显式生成中间步骤的情况下保留推理能力?

Zhenyi Shen,Hanqi Yan,Linhai Zhang,Zhanghao Hu,Yali Du,Yulan He
latent-reasoningcot-compressionself-distillationDOIDBLP
7
泛读IndustryEMNLP 2025

DAST: Difficulty-Adaptive Slow-Thinking for Large Reasoning Models

大型推理模型(如 o1/R1 类)在所有问题上都使用长推理链(slow-thinking),但简单问题不需要这么多计算,造成资源浪费。本文要解决的是如何根据问题难度自适应地调整推理深度。

Yi Shen,Jian Zhang,Jieyun Huang,Shuming Shi,Wenjing Zhang,Jiangze Yan,Ning Wang,Kai Wang,Zhaoxiang Liu,Shiguo Lian
slow-thinkingtest-time-computereasoningDOIDBLP
7
精读EMNLP 2025

Route Sparse Autoencoder to Interpret Large Language Models

Sparse Autoencoder (SAE) 是当前 LLM 可解释性研究的主流工具,但标准 SAE 对所有输入使用相同的字典,缺乏对不同输入类型的适应性。本文提出用路由机制改进 SAE,使其能更好地解释 LLM 的内部表示。

Wei Shi,Sihang Li,Tao Liang,Mingyang Wan,Guojun Ma,Xiang Wang,Xiangnan He
sparse-autoencoderinterpretabilityroutingDOIDBLP
7
精读EMNLP 2025

Warm Up Before You Train: Unlocking General Reasoning in Resource-Constrained Settings

在资源受限(如小参数量或有限算力)的设定下,直接使用复杂的推理数据从头训练模型,往往会导致模型陷入局部最优,只能记住表面模式而无法涌现出真正的泛化推理能力。

Safal Shrestha,Minwu Kim,Aadim Nepal,Anubhav Shrestha,Keith W. Ross
pretraining-strategyreasoningresource-constrainedDOIDBLP
7
精读EMNLP 2025

Beyond Input Activations: Identifying Influential Latents by Gradient Sparse Autoencoders

传统的稀疏自编码器(SAE)仅基于前向激活值(Activations)进行重建,这会提取出大量方差虽大但对模型最终输出(Loss)毫无因果影响的冗余“死特征”。

Dong Shu,Xuansheng Wu,Haiyan Zhao,Mengnan Du,Ninghao Liu
sparse-autoencoderinterpretabilitylatent-featuresDOIDBLP
7
精读FindingsEMNLP 2025

A Survey on Sparse Autoencoders: Interpreting the Internal Mechanisms of Large Language Models

随着稀疏自编码器(SAE)在 LLM 机制可解释性(Mechanistic Interpretability)领域的爆发式增长,文献中充斥着各种架构变体、训练目标和评估指标,缺乏一个统一的分类学框架。

Dong Shu,Xuansheng Wu,Haiyan Zhao,Daking Rai,Ziyu Yao,Ninghao Liu,Mengnan Du
sparse-autoencoderinterpretabilitysurveyAnthologyDBLP
7
精读EMNLP 2025

GATEAU: Selecting Influential Samples for Long Context Alignment

长上下文(Long-context)SFT 数据构建和训练成本极高,且随机采样长文档容易引入大量冗余噪声,反而稀释了模型对关键信息的注意力,导致对齐效果次优。

Shuzheng Si,Haozhe Zhao,Gang Chen,Yunshui Li,Kangyang Luo,Chuancheng Lv,Kaikai An,Fanchao Qi,Baobao Chang,Maosong Sun
long-contextdata-selectionalignmentDOIDBLP
7
精读EMNLP 2025

(Almost) Free Modality Stitching of Foundation Models

这篇工作要解决的是:能否在几乎不重新训练基础模型的前提下,把不同模态的现成 foundation model 低成本拼接成一个可协同工作的系统。过去常见做法要么做大规模联合预训练,要么靠重型对齐模块,成本高且容易破坏原模型能力;因此“接近免费”的 modality stitching 如果成立,对统一多模态预训练和系统组装都有直接价值。

Jaisidh Singh,Diganta Misra,Boris Knyazev,Antonio Orvieto
modality-stitchingfoundation-modelsmultimodal-architectureDOIDBLP
7
精读FindingsEMNLP 2025

You Only Use Reactive Attention Slice When Retrieving From Long Context

这篇工作的核心问题是:长上下文检索时,模型真正用到的注意力计算是否远小于我们为整个上下文付出的成本。现有长上下文方法普遍按全局预算做 attention 优化,但很多 token 并不会参与有效检索;因此如果 retrieval 只依赖一个 reactive attention slice,就存在显著的计算与内存浪费空间。

Yun Joon Soh,Hanxian Huang,Yuandong Tian,Jishen Zhao
long-contextattentionretrievalAnthologyDBLP
7
精读EMNLP 2025

Accelerated Test-Time Scaling with Model-Free Speculative Sampling

这篇工作解决的是 test-time scaling 的推理成本问题:如何在保留更强采样或搜索收益的同时,把生成速度拉回来。现有 speculative decoding 往往需要一个 draft model,而这引入额外模型维护和失配问题;model-free speculative sampling 试图不依赖草稿模型做加速。

Woomin Song,Saket Dingliwal,Sai Muralidhar Jayanthi,Bhavana Ganesh,Jinwoo Shin,Aram Galstyan,Sravan Babu Bodapati
speculative-decodingsamplingtest-time-scalingDOIDBLP
7
精读EMNLP 2025

Mixture of Weight-shared Heterogeneous Group Attention Experts for Dynamic Token-wise KV Optimization

Guanghui Song,Dongping Liao,Yiren Zhao,Kejiang Ye,Chengzhong Xu,Xitong Gao
kv-cacheattentionmoeDOIDBLP
7
精读FindingsEMNLP 2025

The Hallucination Tax of Reinforcement Finetuning

Linxin Song,Taiwei Shi,Jieyu Zhao
reinforcement-learninghallucinationalignmentAnthologyDBLP
7
泛读EMNLP 2025

The Role of Outgoing Connection Heterogeneity in Feedforward Layers of Large Language Models

Felix Stahlberg,Shankar Kumar
ffnarchitectureinterpretabilityDOIDBLP
7
泛读FindingsEMNLP 2025

SPO: Self Preference Optimization with Self Regularization

Yuhao Sun,Yifan Zhang,Quandong Wang,Qinzhuo Wu,Wei Liu,Jian Luan
dpoself-preferencealignmentAnthologyDBLP
7
泛读EMNLP 2025

Identifying Pre-training Data in LLMs: A Neuron Activation-Based Detection Framework

Hongyi Tang,Zhihao Zhu,Yi Yang
pretraining-datamembership-inferenceneuron-analysisDOIDBLP
7
泛读EMNLP 2025

The Transfer Neurons Hypothesis: An Underlying Mechanism for Language Latent Space Transitions in Multilingual LLMs

Hinata Tezuka,Naoya Inoue
multilingualinterpretabilityneuron-analysisDOIDBLP
7
泛读FindingsEMNLP 2025

Breaking Token Into Concepts: Exploring Extreme Compression in Token Representation Via Compositional Shared Semantics

标准的 Token Embedding 层参数量庞大(尤其在多语言或大词表下),且表示高度冗余,缺乏组合语义结构。

Kavin R. V.,Pawan Goyal
IIT Kharagpurtokenizertoken-representationcompressionAnthologyDBLP
7
泛读EMNLP 2025

Context-aware Biases for Length Extrapolation

这篇论文要解决的是:Transformer 做长度外推时,位置编码之外还存在被忽视的上下文依赖偏置,导致模型在训练长度外快速失稳。过去大多数工作把问题几乎都归因于 RoPE、ALiBi 或 attention score 的位置设计,因此主要改位置编码;这篇工作认为,真正限制外推的还有由上下文统计诱导出的偏置项,值得单独建模和校正。

Ali Veisi,Hamidreza Amirzadeh,Amir Mansourian
length-extrapolationpositional-encodinglong-contextDOIDBLP
7
泛读EMNLP 2025

SliceMoE: Routing Embedding Slices Instead of Tokens for Fine-Grained and Balanced Transformer Scaling

这篇论文要解决的是:MoE Transformer 的路由粒度过粗,按 token 路由容易造成专家负载不均,也限制了细粒度容量利用。现有 MoE 大多默认 token 是最自然的路由单位,但一个 token embedding 内部不同子空间可能对应不同功能,把整 token 送给一个专家会浪费可分解结构。

Harshil Vejendla
moeroutingfine-grained-moeDOIDBLP
7
泛读FindingsEMNLP 2025

Train a Unified Multimodal Data Quality Classifier with Synthetic Data

这篇论文要解决的是:多模态数据质量筛选越来越重要,但人工标注一个统一的数据质量分类器成本很高、覆盖也差。过去常见做法是按模态分别写规则或训练专用过滤器,这会导致标准不一致,也很难扩展到新模态或新噪声类型。

Weizhi Wang,Rongmei Lin,Shiyang Li,Colin Lockard,Ritesh Sarkhel,Sanket Lokegaonkar,Jingbo Shang,Xifeng Yan,Nasser Zalmout,Xian Li
data-qualitysynthetic-datamultimodalAnthologyDBLP
7
精读FindingsEMNLP 2025

LongLLaVA: Scaling Multi-modal LLMs to 1000 Images Efficiently via a Hybrid Architecture

Xidong Wang,Dingjie Song,Shunian Chen,Junying Chen,Zhenyang Cai,Chen Zhang,Lichao Sun,Benyou Wang
multimodallong-contexthybrid-architectureAnthologyDBLP
7
泛读FindingsEMNLP 2025

Benchmarking Contextual and Paralinguistic Reasoning in Speech-LLMs: A Case Study with In-the-Wild Data

Qiongqiong Wang,Hardik B. Sailor,Tianchi Liu,Wenyu Zhang,Muhammad Huzaifah,Nattadaporn Lertcheva,Shuo Sun,Nancy F. Chen,Jinyang Wu,AiTi Aw
speech-lmbenchmarkparalinguisticAnthologyDBLP
7
泛读EMNLP 2025

Language Models as Continuous Self-Evolving Data Engineers

Peidong Wang,Ming Wang,Zhiming Ma,Xiaocui Yang,Shi Feng,Daling Wang,Yifei Zhang,Kaisong Song
self-evolvingdata-synthesiscontinual-pretrainDOIDBLP
7
泛读FindingsEMNLP 2025

NeighXLM: Enhancing Cross-Lingual Transfer in Low-Resource Languages via Neighbor-Augmented Contrastive Pretraining

Sicheng Wang,Wenyi Wu,Zibo Zhang
multilinguallow-resourcecontrastive-pretrainAnthologyDBLP
7
泛读FindingsEMNLP 2025

Understanding the Language Model to Solve the Symbolic Multi-Step Reasoning Problem from the Perspective of Buffer Mechanism

Zhiwei Wang,Yunji Wang,Zhongwang Zhang,Zhangchen Zhou,Hui Jin,Tianyang Hu,Jiacheng Sun,Zhenguo Li,Yaoyu Zhang,Zhi-Qin John Xu
reasoninginterpretabilitybufferAnthologyDBLP
7
泛读FindingsEMNLP 2025

Sparsifying Mamba

An Wang,Ruobing Xie,Shuaipeng Li,Xingwu Sun,Zhanhui Kang
mambassmsparsityAnthologyDBLP
7
泛读EMNLP 2025

Language Model Based Text-to-Audio Generation: Anti-Causally Aligned Collaborative Residual Transformers

Juncheng Wang,Chao Xu,Cheng Yu,Zhe Hu,Haoyu Xie,Guoqi Yu,Lei Shang,Shujun Wang
audio-lmtext-to-audioresidual-transformerDOIDBLP
7
泛读EMNLP 2025

Distribution Prompting: Understanding the Expressivity of Language Models Through the Next-Token Distributions They Can Produce

传统评估 LLM 表达能力的方法多依赖离散 token 输出,难以全面衡量模型在整个词表分布上的拟合与控制能力。

Haojin Wang,Zining Zhu,Freda Shi
University of WaterlooVector Instituteexpressivitynext-tokendistributionDOIDBLP
7
泛读FindingsEMNLP 2025

A Comprehensive Survey on Learning from Rewards for Large Language Models: Reward Models and Learning Strategies

LLM 的 Reward Learning(如 RLHF, DPO, PPO)技术发展迅速,但缺乏对 Reward Models 和学习策略的系统性梳理和对比。

Xiaobao Wu
surveyreward-modelrlhfAnthologyDBLP
7
泛读EMNLP 2025

TokenSelect: Efficient Long-Context Inference and Length Extrapolation for LLMs via Dynamic Token-Level KV Cache Selection

LLM 长上下文推理时,KV Cache 随序列长度线性增长,导致显存溢出和解码延迟。

Wei Wu,Zhuoshi Pan,Kun Fu,Chao Wang,Liyi Chen,Yunchu Bai,Tianfu Wang,Zheng Wang,Hui Xiong
kv-cachelong-contextinferenceDOIDBLP
7
精读EMNLP 2025

Parallel Continuous Chain-of-Thought with Jacobi Iteration

连续思维链(Continuous CoT)通过在隐空间生成 dummy tokens 提升推理能力,但自回归生成这些隐状态会导致严重的推理延迟。

Haoyi Wu,Zhihao Teng,Kewei Tu
ShanghaiTech Universitychain-of-thoughtparallel-decodingjacobi-iterationDOIDBLP
7
泛读EMNLP 2025

Continual Learning of Large Language Models

这篇论文要解决的是:大语言模型在持续学习中如何吸收新任务或新域,同时尽量不遗忘旧能力。这个问题过去常被用全量重训、混合回放或参数隔离来绕开,但这些办法要么算力和数据代价高,要么会压缩模型的可塑性,因此在真实的长期演化场景里仍然不够好。

Tongtong Wu,Thuy-Trang Vu,Linhao Luo,Gholamreza Haffari
continual-learningcontinual-pretrainsurveyDOIDBLP
7
泛读EMNLP 2025

Enhancing LLM Language Adaption through Cross-lingual In-Context Pre-training

这篇论文要解决的是:如何让 LLM 更高效地适应新语言,而不是只靠目标语言继续预训练或翻译式数据扩增。过去的语言适配通常把跨语言迁移当成静态数据覆盖问题处理,但这会低估上下文内跨语言示例对推理模式迁移的作用。

Linjuan Wu,Haoran Wei,Huan Lin,Tianhao Li,Baosong Yang,Fei Huang,Weiming Lu
cross-lingualin-context-pretrainlanguage-adaptationDOIDBLP
7
精读EMNLP 2025

Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge

这篇论文要解决的是:能否用 LLM 作为 meta-judge 来持续改进另一个语言模型的对齐,而不是依赖固定的人类偏好数据或静态 reward model。过去的自举式对齐常受限于 judge 偏差会被放大,一旦奖励器不可靠,就会把错误偏好持续写回模型。

Tianhao Wu,Weizhe Yuan,Olga Golovneva,Jing Xu,Yuandong Tian,Jiantao Jiao,Jason E. Weston,Sainbayar Sukhbaatar
self-improvingreward-modelmeta-judgeDOIDBLP
7
泛读EMNLP 2025

Understanding the Modality Gap: An Empirical Study on the Speech-Text Alignment Mechanism of Large Speech Language Models

这篇论文要解决的是:大型语音语言模型里 speech-text alignment 的机制到底是什么,所谓 modality gap 具体表现在哪里。过去很多工作展示了语音接入 LLM 后任务可做,但对对齐是如何形成的、失败点在表示层还是优化层,分析还不够细,这限制了 tokenizer、连接器和训练目标的系统设计。

Bajian Xiang,Shuaijiang Zhao,Tingwei Guo,Wei Zou
speech-lmmodality-gapalignment-mechanismDOIDBLP
7
泛读EMNLP 2025

Audio-Reasoner: Improving Reasoning Capability in Large Audio Language Models

大型音频语言模型(LALM)在感知类任务上还行,但面对需要多步推理的音频问题(声学场景分析、复杂事件因果、音乐/对话理解)表现明显差于文本 LLM 在文本上的推理水平。作者想把 CoT-style 推理引入 LALM。

Zhifei Xie,Mingbao Lin,Zihang Liu,Pengcheng Wu,Shuicheng Yan,Chunyan Miao
audio-lmreasoningmultimodalDOIDBLP
7
泛读EMNLP 2025

The Emperor's New Reasoning: Format Imitation Overshadows Genuine Mathematical Understanding in SFT

在数学推理任务上,使用 CoT 数据进行 SFT 往往能带来指标提升,但这究竟是因为模型真正学会了数学逻辑,还是仅仅学会了模仿 CoT 的“步骤格式”?

Linyao Yang,Jian-Tao Huang,Yafei Lu,Zhenhui Jessie Li,Guirong Xue
sftreasoningformat-imitationDOIDBLP
7
泛读EMNLP 2025

Internal Chain-of-Thought: Empirical Evidence for Layer-wise Subtask Scheduling in LLMs

LLM 在不输出显式思维链(CoT)时,是如何在内部完成复杂任务的?其深层 Transformer 结构中是否存在隐式的、逐层递进的子任务调度机制?

Zhipeng Yang,Junzhuo Li,Siyu Xia,Xuming Hu
cotinterpretabilitylayer-wiseDOIDBLP
7
泛读FindingsEMNLP 2025

Learning Is Not A Race: Improving Retrieval in Language Models via Equal Learning

在检索增强(RAG)或基于检索的语言模型训练中,模型倾向于走捷径——过度依赖容易学习的检索信号(如表面词汇重叠),而忽略更难但信息量更大的深层上下文,导致优化过程中的“学习不均衡(Race)”。

Wanqian Yang,Aahlad Manas Puli,Rajesh Ranganath
New York University (NYU)retrievaltraining-dynamicslearning-rateAnthologyDBLP
7
泛读EMNLP 2025

How Does DPO Reduce Toxicity? A Mechanistic Neuron-Level Analysis

直接偏好优化(DPO)在降低模型毒性(Toxicity)方面非常有效,但它在底层神经元级别究竟改变了什么?是“遗忘”了有毒概念,还是仅仅“抑制”了有毒输出?

Yushi Yang,Filip Sondej,Harry Mayne,Andrew Lee,Adam Mahdi
dpointerpretabilitytoxicityDOIDBLP
7
泛读EMNLP 2025

Unveiling Internal Reasoning Modes in LLMs: A Deep Dive into Latent Reasoning vs. Factual Shortcuts with Attribute Rate Ratio

在评估 LLM 的推理能力时,很难区分模型是真的在内部进行了逻辑推导(Latent Reasoning),还是仅仅依靠预训练记忆走捷径(Factual Shortcuts,即直接背出了答案)。

Yiran Yang,Haifeng Sun,Jingyu Wang,Qi Qi,Zirui Zhuang,Huazheng Wang,Pengfei Ren,Jing Wang,Jianxin Liao
latent-reasoninginterpretabilityfactualDOIDBLP
7
泛读EMNLP 2025

DCIS: Efficient Length Extrapolation of LLMs via Divide-and-Conquer Scaling Factor Search

这篇工作要解决的是:在不重新训练模型的前提下,如何更便宜地为长上下文外推找到合适的 RoPE/位置缩放因子。现有做法通常靠网格搜索、经验公式或少量任务上手调,成本高且常常对模型、长度区间和任务分布不稳,因此这个问题对长上下文部署和 mid-train 调参都很实际。

Lei Yang,Shaoyang Xu,Jianxiang Peng,Shaolin Zhu,Deyi Xiong
long-contextlength-extrapolationropeDOIDBLP
7
泛读FindingsEMNLP 2025

Unearthing Gems from Stones: Policy Optimization with Negative Sample Augmentation for LLM Reasoning

这篇工作要解决的是:推理类 RL/策略优化里,负样本通常只被当作被淘汰的轨迹,而没有被充分利用来塑造更强的学习信号。过去很多方法主要放大正样本或高回报轨迹,但在难推理任务里,正样本稀缺,单靠它们会让学习信号又稀又偏。

Zhaohui Yang,Yuxiao Ye,Shilei Jiang,Shihong Deng,Chen Hu,Linjing Li,Daxin Jiang
rlreasoningnegative-samplesAnthologyDBLP
7
泛读EMNLP 2025

XQuant: Achieving Ultra-Low Bit KV Cache Quantization with Cross-Layer Compression

这篇工作要解决的是:KV cache 在长上下文推理中占用显存过大,而极低比特量化通常会明显伤害生成质量。传统 KV 量化大多逐层独立处理,忽略了层间冗余,因此在 2-bit、甚至更低位宽时误差很难压住。

Haoqi Yang,Yao Yao,Zuchao Li,Baoyuan Qi,Guoming Liu,Hai Zhao
kv-cachequantizationcross-layerDOIDBLP
7
泛读EMNLP 2025

A Probabilistic Inference Scaling Theory for LLM Self-Correction

这篇工作要解决的是:LLM 自纠错在什么条件下有效、什么时候只是多做几轮无用推理,目前缺少一个可检验的理论框架。过去大量工作把 self-correction 当经验技巧在用,但对其 scaling 规律、收益边界和失败模式解释不足。

Zhe Yang,Yichang Zhang,Yudong Wang,Ziyao Xu,Junyang Lin,Zhifang Sui
inference-scalingself-correctiontheoryDOIDBLP
7
泛读EMNLP 2025

Language models can learn implicit multi-hop reasoning, but only if they have lots of training data

这篇工作要解决的是:语言模型是否能从纯序列建模中学到隐式 multi-hop reasoning,以及这种能力需要什么数据条件。过去很多工作在能力层面给出正反两种结论,但往往混杂了模型大小、任务模板和训练分布,导致大家不清楚限制到底来自架构、目标函数还是数据量。

Yuekun Yao,Yupei Du,Dawei Zhu,Michael Hahn,Alexander Koller
multi-hop-reasoningtraining-datascalingDOIDBLP
7
泛读EMNLP 2025

Scalable Data Synthesis through Human-like Cognitive Imitation and Data Recombination

Zhongyi Ye,Weitai Zhang,Xinyuan Zhou,Yongxin Zhu,Ninghui Rao,Enhong Chen
data-synthesisscalingDOIDBLP
7
泛读FindingsEMNLP 2025

MANTA: A Scalable Pipeline for Transmuting Massive Web Corpora into Instruction Datasets

Heuiyeen Yeen,Seokhee Hong,Hyeongu Yun,Jinsik Lee
instruction-dataweb-corporadata-synthesisAnthologyDBLP
7
泛读FindingsEMNLP 2025

Understanding Refusal in Language Models with Sparse Autoencoders

Wei Jie Yeo,Nirmalendu Prakash,Clement Neo,Ranjan Satapathy,Roy Ka-Wei Lee,Erik Cambria
saerefusalinterpretabilityAnthologyDBLP
7
精读EMNLP 2025

Accelerate Parallelizable Reasoning via Parallel Decoding within One Sequence

Yijiong Yu,Wei Wang,Ran Chen,Ji Pei
parallel-decodingreasoninginference-efficiencyDOIDBLP
7
精读FindingsEMNLP 2025

ReMamba: Equip Mamba with Effective Long-Sequence Modeling

Danlong Yuan,Jiahao Liu,Bei Li,Huishuai Zhang,Jingang Wang,Xunliang Cai,Dongyan Zhao
mambassmlong-contextAnthologyDBLP
7
泛读EMNLP 2025

Enhancing Large Vision-Language Models with Ultra-Detailed Image Caption Generation

Yu Zeng,Yukun Qi,Yiming Zhao,Xikun Bao,Lin Chen,Zehui Chen,Shiting Huang,Jie Zhao,Feng Zhao
vlmdata-qualityimage-captioningDOIDBLP
7
泛读EMNLP 2025

Thinking Out Loud: Do Reasoning Models Know When They're Right?

Qingcheng Zeng,Weihao Xuan,Leyang Cui,Rob Voigt
reasoningcalibrationcotDOIDBLP
7
泛读EMNLP 2025

BTS: Harmonizing Specialized Experts into a Generalist LLM

Qizhen Zhang,Prajjwal Bhargava,Chloe Bi,Chris X. Cai,Jakob Nicolaus Foerster,Jeremy Fu ... 省略 2 位作者 ... ,Sheng Shen,Emily Dinan,Suchin Gururangan,Mike Lewis
moemodel-merginggeneralist-modelDOIDBLP
7
泛读EMNLP 2025

Direct Value Optimization: Improving Chain-of-Thought Reasoning in LLMs with Refined Values

Hongbo Zhang,Han Cui,Guangsheng Bao,Linyi Yang,Jun Wang,Yue Zhang
reasoningcotrlhfDOIDBLP
7
泛读EMNLP 2025

Complex Numerical Reasoning with Numerical Semantic Pre-training Framework

Jun Zhang,Haihong E,Tianyi Hu,Yifan Zhu,Meina Song,Haoran Luo
pretrainingnumerical-reasoningmathDOIDBLP
7
泛读EMNLP 2025

Beyond Online Sampling: Bridging Offline-to-Online Alignment via Dynamic Data Transformation for LLMs

在线 RLHF(如 PPO)采样成本高且训练不稳定,而离线对齐(如 DPO)由于静态数据集与模型当前策略存在分布偏移(OOD),难以达到在线方法的上限。

Zhang Zhang,Guhao Feng,Jian Guan,Di He,Wei Wu
Peking Universityalignmentrlhfoffline-to-onlineDOIDBLP
7
泛读EMNLP 2025

AnchorAttention: Difference-Aware Sparse Attention with Stripe Granularity

长上下文 LLM 的全注意力机制计算复杂度为 O(N^2),而现有的静态稀疏注意力(如局部+步长)无法捕捉动态的长程依赖,且对硬件内存访问不友好。

Yu Zhang,Dong Guo,Fang Wu,Guoliang Zhu,Dian Ding,Yiming Zhang
attentionsparse-attentionlong-contextDOIDBLP
7
精读EMNLP 2025

From Reasoning to Answer: Empirical, Attention-Based and Mechanistic Insights into Distilled DeepSeek R1 Models

这篇工作聚焦一个现在很关键但常被模糊处理的问题:DeepSeek R1 蒸馏模型到底是“学会了推理”,还是只学会了把推理轨迹压缩成答案。此前很多分析停留在 benchmark 分数或 CoT 长度统计,缺少对 attention 行为和内部机制的联合证据,因此这个问题值得重新拆开看。

Jue Zhang,Qingwei Lin,Saravan Rajmohan,Dongmei Zhang
reasoningmechanistic-interpretabilityattentionDOIDBLP
7
精读EMNLP 2025

Merge then Realign: Simple and Effective Modality-Incremental Continual Learning for Multimodal LLMs

这篇工作解决的是 multimodal LLM 在 modality-incremental continual learning 里的核心矛盾:新模态接进来时,模型既要快速吸收新能力,又不能破坏已有模态对齐。现有方法常在保留旧能力和学习新模态之间拉扯,或者需要复杂回放与正则,工程代价高。

Dingkun Zhang,Shuhan Qi,Xinyu Xiao,Kehai Chen,Xuan Wang
continual-learningmultimodal-llmmodality-incrementalDOIDBLP
7
泛读EMNLP 2025

Speculating LLMs' Chinese Training Data Pollution from Their Tokens

中文 LLM 的 tokenizer 词表里混入了大量可疑 token——色情、赌博、垃圾站点的片段。作者想把这些 token 当成训练数据污染的指纹,倒推主流 LLM 的中文预训练语料里到底混了多少脏数据。

Qingjie Zhang,Di Wang,Haoting Qian,Liu Yan,Tianwei Zhang,Ke Xu,Qi Li,Minlie Huang,Hewu Li,Han Qiu
tokenizerdata-qualitychineseDOIDBLP
7
泛读EMNLP 2025

Draft Model Knows When to Stop: Self-Verification Speculative Decoding for Long-Form Generation

长文本生成下的投机解码(speculative decoding)效率不稳定:draft 模型在某些片段 accept 率高,在另一些片段低,却还在白跑 K 步 draft。作者让 draft 模型自己判断什么时候该停。

Ziyin Zhang,Jiahao Xu,Tian Liang,Xingyu Chen,Zhiwei He,Rui Wang,Zhaopeng Tu
speculative-decodinglong-generationself-verificationDOIDBLP
7
泛读EMNLP 2025

Query-Focused Retrieval Heads Improve Long-Context Reasoning and Re-ranking

Wuwei Zhang,Fangcong Yin,Howard Yen,Danqi Chen,Xi Ye
long-contextretrieval-headsattentionDOIDBLP
7
泛读EMNLP 2025

LightThinker: Thinking Step-by-Step Compression

Jintian Zhang,Yuqi Zhu,Mengshu Sun,Yujie Luo,Shuofei Qiao,Lun Du,Da Zheng,Huajun Chen,Ningyu Zhang
cot-compressionreasoningkv-cacheDOIDBLP
7
泛读IndustryEMNLP 2025

More Data or Better Data? A Critical Analysis of Data Selection and Synthesis for Mathematical Reasoning

数学推理的 SFT/posttrain 阶段,到底是更多数据有效还是更好的数据有效?现有工作在数据选择和数据合成两个方向上各有进展,但缺乏系统性的对比分析来回答这个根本问题。

Yike Zhao,Simin Guo,Ziqing Yang,Shifan Han,Dahua Lin,Fei Tan
data-selectionmath-reasoningdata-synthesisDOIDBLP
7
泛读EMNLP 2025

QSpec: Speculative Decoding with Complementary Quantization Schemes

在推测解码(Speculative Decoding)中,为了极致加速通常会对草稿模型和目标模型进行量化,但独立量化会导致两者输出分布偏移,严重降低接受率(Acceptance Rate)。

Juntao Zhao,Wenhao Lu,Sheng Wang,Lingpeng Kong,Chuan Wu
HKUspeculative-decodingquantizationDOIDBLP
7
泛读EMNLP 2025

MuTIS: Enhancing Reasoning Efficiency through Multi Turn Intervention Sampling in Reinforcement Learning

在针对推理任务的强化学习中,由于推理步骤长且最终奖励稀疏,传统的序列级采样效率极低,模型很难探索到正确的推理路径。

Wenshuo Zhao,Haoxing Zhai,Xinyu Qiu,Zhenting Qi,Shuhe Li,Linchao Zhu
rlreasoningefficiencyDOIDBLP
7
泛读EMNLP 2025

StepSearch: Igniting LLMs Search Ability via Step-Wise Proximal Policy Optimization

这篇论文要解决的是:现有 LLM 搜索能力训练通常把整轮搜索当作单一步决策来优化,导致奖励稀疏、信用分配粗糙,模型学不会稳定的多步检索与证据整合。这个问题过去常被工具调用模板、蒸馏或结果级 RL 暂时绕开,但一旦进入真实搜索环境,错误会在中间步骤迅速累积,所以值得单独做 step-wise 优化。

Xuhui Zheng,Kang An,Ziliang Wang,Yuhang Wang,Yichao Wu
rlsearchppoDOIDBLP
7
泛读EMNLP 2025

DeepResearcher: Scaling Deep Research via Reinforcement Learning in Real-world Environments

这篇论文要解决的是:如何让 LLM 在真实环境里完成“深度研究”这类长周期、高外部依赖的任务,而不是只在静态 benchmark 上做短链推理。过去很多工作在合成环境或离线问答上训练 agent,规避了真实网页、延迟、噪声和工具失败;但真正影响实用性的正是这些环境因素。

Yuxiang Zheng,Dayuan Fu,Xiangkun Hu,Xiaojie Cai,Lyumanshan Ye,Pengrui Lu,Pengfei Liu
rldeep-researchagentDOIDBLP
7
泛读FindingsEMNLP 2025

Toward Optimal LLM Alignments Using Two-Player Games

这篇论文要解决的是:现有 LLM 对齐大多把训练写成单边优化,例如让策略去贴近偏好数据或奖励模型,但这会忽略策略与评估者之间的相互作用,容易出现过拟合奖励、模式坍缩或局部最优。这个问题过去靠更好的 preference data 或更稳的 objective 缓解,但并没有从博弈结构上重写对齐问题。

Rui Zheng,Hongyi Guo,Zhihan Liu,Xiaoying Zhang,Yuanshun Yao,Xiaojun Xu ... 省略 3 位作者 ... ,Qi Zhang,Xuanjing Huang,Yang Liu,Hang Li
alignmentrlself-playAnthologyDBLP
7
泛读FindingsEMNLP 2025

DynamicKV: Task-Aware Adaptive KV Cache Compression for Long Context LLMs

这篇论文要解决的是:长上下文 LLM 的 KV cache 压缩通常按通用启发式做,无法针对具体任务保留真正有用的历史 token,结果要么省不了多少显存,要么损失回答质量。过去很多方法按注意力分数、位置或固定预算压缩,但没有充分利用‘不同任务需要不同记忆’这一事实。

Xiabin Zhou,Wenbin Wang,Minyan Zeng,Jiaxian Guo,Xuebo Liu,Li Shen,Min Zhang,Liang Ding
kv-cachelong-contextcompressionAnthologyDBLP
7
泛读FindingsEMNLP 2025

Dropping Experts, Recombining Neurons: Retraining-Free Pruning for Sparse Mixture-of-Experts LLMs

Yixiao Zhou,Ziyu Zhao,Dongzhou Cheng,Zhiliang Wu,Jie Gui,Yi Yang,Fei Wu,Yu Cheng,Hehe Fan
moepruningretraining-freeAnthologyDBLP
6
泛读EMNLP 2025

NormXLogit: The Head-on-Top Never Lies

这篇论文要解决的是:怎样用一个模型无关、计算便宜的方法衡量输入 token 对 LLM 输出的重要性,而不依赖昂贵且模型特定的解释性工具。现有可解释方法常常需要访问内部梯度、特定层结构或反事实重算,迁移性和成本都不理想;作者试图直接利用 embedding 与输出头之间的几何关系给出 token 重要性信号。

Sina Abbasi,Mohammad Reza Modarres,Mohammad Taher Pilehvar
interpretabilitylogit-lensDOIarXivDBLP
6
泛读FindingsEMNLP 2025

SAFE: A Sparse Autoencoder-Based Framework for Robust Query Enrichment and Hallucination Mitigation in LLMs

这篇论文要解决的是:LLM 幻觉不仅要检测,还要在检索或问答前就主动改写查询以减少后续幻觉,但现有工作通常把“检测”和“缓解”分开做,且很少利用 SAE 提供的可解释内部特征。过去方案往往在生成后做事实核验或置信过滤,能发现问题,但很难在源头上改变模型会去检索什么、会如何组织问题。

Samir Abdaljalil,Filippo Pallucchini,Andrea Seveso,Hasan Kurban,Fabio Mercorio,Erchin Serpedin
saehallucinationinterpretabilityAnthologyarXivDBLP
6
泛读EMNLP 2025

Mechanisms vs. Outcomes: Probing for Syntax Fails to Explain Performance on Targeted Syntactic Evaluations

这篇论文要解决的是:线性 probing 测到的“句法信息可读出”到底能不能解释模型在真实句法测试中的表现。过去很多工作默认把 probe accuracy 当作模型内部掌握句法机制的证据,但这一步经常把“能从表示里读出来”混同为“模型实际依赖它完成任务”。作者要检验的正是这种机制—结果之间是否真的对应。

Ananth Agarwal,Jasper Jian,Christopher D. Manning,Shikhar Murty
Stanford UniversityprobingsyntaxinterpretabilityDOIarXivDBLP
6
泛读FindingsEMNLP 2025

KurTail : Kurtosis-based LLM Quantization

这篇论文要解决的是:LLM 在 4-bit 量化时最棘手的问题仍是激活和 KV cache 中的 outlier,常规均匀量化在这种长尾分布下误差过大。已有方法会做旋转或校正,但对 outlier 的处理还不够直接,尤其很难同时把权重、激活和 KV cache 都压到 4 bit 而不显著伤精度。

Mohammad Sadegh Akhondzadeh,Aleksandar Bojchevski,Evangelos Eleftheriou,Martino Dazzi
quantizationoutliersAnthologyarXivDBLP
6
泛读EMNLP 2025

TRACE: Training and Inference-Time Interpretability Analysis for Language Models

现有 interpretability 工具大多是 post-hoc、标量指标为主、集成成本高,难以在训练过程中连续观察语言知识的涌现。作者想要一个能随训练动态跑的分析工具链。

Nura Aljaafari,Danilo S. Carvalho,André Freitas
University of ManchesterIdiap Research Instituteinterpretabilitytraining-dynamicsprobingDOIarXivDBLP
6
泛读FindingsEMNLP 2025

Revealing the Inherent Instructability of Pre-Trained Language Models

预训练 base model 通常被认为需要 SFT 才能跟随指令;作者主张 base model 本身就具备可指令化的能力,只是没被正确激发。

Seokhyun An,Minji Kim,Hyounghun Kim
instruction-followingpretrained-lmemergent-abilityAnthologyDBLP
6
泛读EMNLP 2025

UltraIF: Advancing Instruction Following from the Wild

开源模型在复杂指令跟随(多约束、嵌套约束)上和闭源差距明显。核心困难是:真实复杂指令怎么 scale 合成?人工写贵,纯模型合成又容易 collapse。

Kaikai An,Li Sheng,Ganqu Cui,Shuzheng Si,Ning Ding,Yu Cheng,Baobao Chang
Tsinghua UniversityShanghai AI Laboratoryinstruction-followingsftdata-synthesisDOIarXivDBLP
6
泛读EMNLP 2025

Long Chain-of-Thought Fine-tuning via Understanding-to-Reasoning Transition

长链 CoT SFT(蒸馏 o1/R1 风格推理轨迹)常出现模型没真正'理解'中间步骤就硬背格式的问题。作者关注如何在 SFT 阶段让模型从'理解问题'平滑过渡到'进行推理'。

Chenxin An,Zhihui Xie,Xiaonan Li,Ming Zhong,Shansan Gong,Lei Li,Jun Zhang,Jingjing Xu,Lingpeng Kong
The University of Hong KongShanghai AI Laboratorylong-cotfine-tuningreasoningDOIDBLP
6
泛读EMNLP 2025

Understanding and Leveraging the Expert Specialization of Context Faithfulness in Mixture-of-Experts LLMs

LLM 在长上下文或 RAG 场景中容易产生幻觉(忽略给定上下文)。在 MoE 架构中,尚不清楚这种“上下文忠实度”是否由特定的专家网络(Experts)负责。

Jun Bai,Minghao Tong,Yang Liu,Zixia Jia,Zilong Zheng
moeexpert-specializationcontext-faithfulnessDOIarXivDBLP
6
泛读EMNLP 2025

Train One Sparse Autoencoder Across Multiple Sparsity Budgets to Preserve Interpretability and Accuracy

传统的稀疏自编码器(SAE)在训练时必须固定稀疏度(k值),为了寻找最佳的解释性分辨率,研究者需要训练和维护多个不同稀疏度的 SAE,计算成本极高。

Nikita Balagansky,Yaroslav Aksenov,Daniil Laptev,Vadim Kurochkin,Gleb Gerasimov,Nikita Koriagin,Daniil Gavrilov
sparse-autoencoderinterpretabilityrepresentation-learningDOIarXivDBLP
6
泛读FindingsEMNLP 2025

Soteria: Language-Specific Functional Parameter Steering for Multilingual Safety Alignment

多语言 LLM 在低资源语言中极易遭遇安全越狱,因为传统的安全对齐(RLHF/SFT)高度依赖英语数据,难以泛化到其他语言。

Somnath Banerjee,Sayan Layek,Pratyush Chatterjee,Animesh Mukherjee,Rima Hazra
safety-alignmentparameter-steeringmultilingualAnthologyarXivDBLP
6
泛读EMNLP 2025

Date Fragments: A Hidden Bottleneck of Tokenization for Temporal Reasoning

Tokenizer对日期字符串的碎片化切分(如将年份切为多个无意义片段)破坏了模型对时间顺序的内在理解,这是导致LLM时间推理能力受损的隐蔽瓶颈。

Gagan Bhatia,Maxime Peyrard,Wei Zhao
tokenizertemporal-reasoningdata-representationDOIDBLP
6
泛读EMNLP 2025

Speculative Streaming: Efficient and Scalable Speculative Decoding with Multi-Stream Attention

传统推测解码依赖外部草稿模型,导致系统复杂度和显存开销增加,难以在资源受限场景下扩展。

Nikhil Bhendawade,Irina Belousova,Qichen Fu,Henry Mason,Antonie Lin,Mohammad Rastegari,Mahyar Najibi
Applespeculative-decodinginference-efficiencyattention-mechanismDOIDBLP
6
泛读EMNLP 2025

Alignment Quality Index (AQI) : Beyond Refusals: AQI as an Intrinsic Alignment Diagnostic via Latent Geometry, Cluster Divergence, and Layer wise Pooled Representations

现有对齐评估过度依赖“拒绝率”,无法区分模型是真正理解了安全边界还是发生了模式崩溃。

Abhilekh Borah,Chhavi Sharma,Danush Khanna,Utkarsh Bhatt,Gurpreet Singh,Hasnat Md Abdullah ... 省略 5 位作者 ... ,Arpita Vats,Rahul Raja,Aman Chadha,Amitava Das
alignmentrepresentationdiagnosticsDOIDBLP
6
泛读FindingsEMNLP 2025

Brittle Minds, Fixable Activations: Understanding Belief Representations in Language Models

LLM内部的事实认知表示脆弱且不一致,导致面对不同提示词时输出矛盾。

Matteo Bortoletto,Constantin Ruhdorfer,Lei Shi,Andreas Bulling
University of Stuttgartbelief-representationactivationinterpretabilityAnthologyDBLP
6
泛读EMNLP 2025

Beyond the Leaderboard: Understanding Performance Disparities in Large Language Models via Model Diffing

仅看Leaderboard总分无法揭示不同LLM在细粒度能力上的差异,难以解释分数相近模型在实际表现上的不同。

Sabri Boughorbel,Fahim Dalvi,Nadir Durrani,Majd Hawasly
Qatar Computing Research Institutemodel-diffingrepresentationevaluationDOIDBLP
6
泛读EMNLP 2025

Beyond Demonstrations: Dynamic Vector Construction from Latent Representations

传统的 In-Context Learning 依赖在 Context 中拼接大量文本示例,推理效率低且容易受 Prompt 表面形式扰动。

Wang Cai,Hsiu-Yuan Huang,Zhixiang Wang,Yunfang Wu
iclrepresentationlatentDOIDBLP
6
泛读FindingsEMNLP 2025

Low-Confidence Gold: Refining Low-Confidence Samples for Efficient Instruction Tuning

在指令微调数据筛选中,低置信度的样本通常被直接丢弃,导致数据浪费且可能丢失困难样本的训练信号。

Hongyi Cai,Jie Li,Mohammad Mahdinur Rahman,Wenzhen Dong
instruction-tuningdata-qualitysample-selectionAnthologyDBLP
6
泛读FindingsEMNLP 2025

PIP: Perturbation-based Iterative Pruning for Large Language Models

对千亿级 LLM 进行剪枝时,一次性剪枝会导致性能断崖式下降,而传统的迭代剪枝计算成本过高。

Yi Cao,Wei-Jie Xu,Yucheng Shen,Weijie Shi,Chi-Min Chan,Jianfeng Qu,Jiajie Xu
Tsinghua UniversityZhipu AIpruningcompressionefficiencyAnthologyDBLP
6
泛读EMNLP 2025

SYNC: A Synthetic Long-Context Understanding Benchmark for Controlled Comparisons of Model Capabilities

现有的长文本 Benchmark 往往将“上下文长度”与“任务复杂度”耦合,导致无法准确归因模型长文本能力的瓶颈。

Shuyang Cao,Kaijian Zou,Lu Wang
University of Michiganlong-contextbenchmarkevaluationDOIDBLP
6
泛读EMNLP 2025

Efficient Beam Search for Large Language Models Using Trie-Based Decoding

这篇工作要解决的是:LLM 上的 beam search 在约束输出、结构化生成或大词表重复前缀场景下会做大量重复计算,导致它相比贪心/采样更难扩展。作者试图用 trie-based decoding 把 beam 间共享前缀显式利用起来,从而降低 beam search 的无效开销。

Brian J. Chan,Mao Xun Huang,Jui-Hung Cheng,Chao-Ting Chen,Hen-Hsen Huang
beam-searchdecodinginferenceDOIDBLP
6
泛读EMNLP 2025

Mind the Gap: How BabyLMs Learn Filler-Gap Dependencies

这篇工作要解决的是:BabyLM 这类小规模语言学习设置能否学到 filler-gap dependency 这种需要长程结构泛化的句法现象,现有工作往往只看整体 benchmark,难以知道模型到底学会了规则还是记住了表面模式。作者把焦点放到一个具体、经典且能区分真正结构泛化与浅层启发式的句法依赖上。

Chi-Yun Chang,Xueyang Huang,Humaira Nasir,Shane Storks,Olawale Akingbade,Huteng Dai
babylmsyntaxlearning-dynamicsDOIDBLP
6
泛读EMNLP 2025

Step-level Verifier-guided Hybrid Test-Time Scaling for Large Language Models

这篇工作要解决的是:test-time scaling 想靠多采样、多步验证或树搜索换性能,但纯生成会浪费算力,纯 verifier 又容易在错误候选上做无效筛选。作者提出 step-level verifier-guided hybrid scaling,目标是在每一步推理时更早地淘汰差路径,把额外计算用在更有希望的推理分支上。

Kaiyan Chang,Yonghao Shi,Chenglong Wang,Hang Zhou,Chi Hu,Xiaoqian Liu,Yingfeng Luo,Yuan Ge,Tong Xiao,JingBo Zhu
test-time-scalingverifierreasoningDOIDBLP
6
泛读EMNLP 2025

Benchmarking Large Language Models Under Data Contamination: A Survey from Static to Dynamic Evaluation

这篇工作要解决的是:LLM 数据污染评测已经从静态 benchmark 泄露演化到更复杂的动态场景,但社区缺少从静态到动态的一体化梳理,导致‘模型到底有没有背题’经常停留在个案讨论。作者把它做成一篇 survey,试图统一污染定义、评测设定和防护思路。

Simin Chen,Yiming Chen,Zexin Li,Yifan Jiang,Zhongwei Wan,Yixin He ... 省略 1 位作者 ... ,Tianle Gu,Haizhou Li,Tao Xie,Baishakhi Ray
contaminationevaluationbenchmarkDOIDBLP
6
泛读FindingsEMNLP 2025

DRBO: Mitigating Short Board Effect via Dynamic Reward Balancing in Multi-reward LLM Optimization

Nuo Chen,Yufei Gao,Yongnan Jin,Yan Hu,Anningzhe Gao,Lingyong Yan,Benyou Wang
rlhfmulti-rewardreward-balancingAnthologyDBLP
6
泛读EMNLP 2025

Icon2: Aligning Large Language Models Using Self-Synthetic Preference Data via Inherent Regulation

Qiyuan Chen,Hongsen Huang,Qian Shao,Jiahe Chen,Jintai Chen,Hongxia Xu,Renjie Hua,Ren Chuan,Jian Wu
alignmentsynthetic-datapreferenceDOIDBLP
6
泛读EMNLP 2025

Multimodal Language Models See Better When They Look Shallower

Haoran Chen,Junyan Lin,Xinghao Chen,Yue Fan,Jianfeng Dong,Xin Jin,Hui Su,Jinlan Fu,Xiaoyu Shen
mllmvision-encoderanalysisDOIDBLP
6
泛读EMNLP 2025

Expanding before Inferring: Enhancing Factuality in Large Language Models through Premature Layers Interpolation

Dingwei Chen,Ziqiang Liu,Feiteng Fang,Chak Tou Leong,Shiwen Ni,Ahmadreza Argha,Hamid Alinejad-Rokny,Min Yang,Chengming Li
factualitylayer-interpolationDOIDBLP
6
泛读FindingsEMNLP 2025

Exploring the Hidden Reasoning Process of Large Language Models by Misleading Them

Guanyu Chen,Peiyang Wang,Yizhou Jiang,Yuqian Liu,Chujie Zhao,Ying Fang,Tianren Zhang,Feng Chen
reasoninginterpretabilityAnthologyDBLP
6
泛读EMNLP 2025

Beyond the Surface: Measuring Self-Preference in LLM Judgments

Zhi-Yuan Chen,Hao Wang,Xinyu Zhang,Enrui Hu,Yankai Lin
llm-judgebiasself-preferenceDOIDBLP
6
泛读FindingsEMNLP 2025

What if Othello-Playing Language Models Could See?

Xinyi Chen,Yifei Yuan,Jiaang Li,Serge J. Belongie,Maarten de Rijke,Anders Søgaard
world-modelprobingmultimodalAnthologyDBLP
6
泛读EMNLP 2025

Extracting and Combining Abilities For Building Multi-lingual Ability-enhanced Large Language Models

Zhipeng Chen,Kun Zhou,Liang Song,Wayne Xin Zhao,Bingning Wang,Weipeng Chen,Ji-Rong Wen
multilingualmodel-mergingcontinual-pretrainDOIDBLP
6
泛读EMNLP 2025

Can LLMs Reason Abstractly Over Math Word Problems Without CoT? Disentangling Abstract Formulation From Arithmetic Computation

Ziling Cheng,Meng Cao,Leila Pishdad,Yanshuai Cao,Jackie CK Cheung
cotreasoningmathDOIDBLP
6
泛读EMNLP 2025

HookMoE: A learnable performance compensation strategy of Mixture-of-Experts for LLM inference acceleration

Longkai Cheng,Along He,Mulin Li,Xueshuo Xie,Tao Li
moeinferenceDOIDBLP
6
泛读FindingsEMNLP 2025

Revisiting Chain-of-Thought Prompting: Zero-shot Can Be Stronger than Few-shot

Zero-shot CoT 在很多场景下被认为弱于 few-shot CoT,但这一结论可能源于 zero-shot prompt 设计不够好。本文重新审视这一对比,探索 zero-shot CoT 能否通过更好的 prompt 策略超越 few-shot CoT。

Xiang Cheng,Chengyan Pan,Minjun Zhao,Deyang Li,Fangchao Liu,Xinyu Zhang,Xiao Zhang,Yong Liu
cotpromptingiclAnthologyDBLP
6
泛读EMNLP 2025

COUNTDOWN: Contextually Sparse Activation Filtering Out Unnecessary Weights in Down Projection

LLM 推理时 FFN 的 down projection 层存在大量冗余激活,如何利用上下文相关的稀疏激活来跳过不必要的计算,从而加速推理。

Jaewon Cheon,Pilsung Kang
sparse-activationinference-efficiencypruningDOIDBLP
6
泛读EMNLP 2025

Toward Efficient Sparse Autoencoder-Guided Steering for Improved In-Context Learning in Large Language Models

Sparse Autoencoder (SAE) 可以提取 LLM 内部的可解释特征,但如何利用这些特征高效地引导(steer)模型行为以改善 in-context learning (ICL) 表现,此前缺乏系统方法。

Ikhyun Cho,Julia Hockenmaier
UIUCsparse-autoencodericlsteeringDOIDBLP
6
泛读FindingsEMNLP 2025

On the Versatility of Sparse Autoencoders for In-Context Learning

SAE 提取的特征在 ICL 中的多功能性(versatility)尚不清楚——这些特征是否能跨任务、跨层地解释和支持 ICL 的不同方面。

Ikhyun Cho,Gaeul Kwon,Julia Hockenmaier
UIUCsparse-autoencodericlinterpretabilityAnthologyDBLP
6
泛读EMNLP 2025

Unsupervised Concept Vector Extraction for Bias Control in LLMs

这篇工作要解决的是:在不依赖人工标注偏见数据的前提下,从 LLM 内部表示里直接抽取可控的“概念向量”,用于干预模型偏见。以往做法通常依赖监督属性标签、人工构造对比样本或外部分类器,因此迁移性差、成本高,也很难覆盖新型偏见维度;如果这篇方法成立,它提供的是一条更轻量的表示层控制路径。

Hannah Cyberey,Yangfeng Ji,David Evans
concept-vectorsbias-controlinterpretabilityDOIDBLP
6
泛读EMNLP 2025

Following the Autoregressive Nature of LLM Embeddings via Compression and Alignment

这篇论文在处理一个很具体但重要的问题:如何让 LLM embedding 更好地遵循语言模型本身的自回归结构。很多 embedding 方法直接把 decoder-only LLM 拿来做 pooling 或对比学习,但句向量空间常与原始 AR 训练目标不一致,导致检索和生成表征之间存在断层。

Jingcheng Deng,Zhongtao Jiang,Liang Pang,Zihao Wei,Liwei Chen,Kun Xu,Yang Song,Huawei Shen,Xueqi Cheng
embeddingsautoregressivecompressionDOIDBLP
6
泛读FindingsEMNLP 2025

Self-Improvement in Multimodal Large Language Models: A Survey

这是一篇 survey,核心问题不是提出新方法,而是系统整理 MLLM self-improvement 的方法谱系、训练信号和评测现状。这个方向最近增长很快,名称相近但机制差异很大;没有统一梳理时,很容易把自训练、反思、合成数据、自奖励和工具反馈混为一谈。

Shijian Deng,Kai Wang,Tianyu Yang,Harsh Singh,Yapeng Tian
surveyself-improvementmllmAnthologyDBLP
6
泛读EMNLP 2025

Theorem-Validated Reverse Chain-of-Thought Problem Generation for Geometric Reasoning

几何推理训练数据稀缺且质量难控:现有合成方法要么生成的题目答案不可验证,要么靠人写模板多样性有限。作者想要一条能自动生成几何题、并且保证答案可证的流水线。

Linger Deng,Linghao Zhu,Yuliang Liu,Yu Wang,Qunyi Xie,Jingjing Wu,Gang Zhang,Yingying Zhu,Xiang Bai
synthetic-datacotreasoningDOIDBLP
6
泛读FindingsEMNLP 2025

Curr-ReFT: Overcoming Training Bottlenecks in Small-scale Vision-Language Models via Curriculum Reinforcement Finetuning

小规模 VLM 直接 RL 微调常常训不动:要么奖励信号太稀疏,要么任务跨度太大导致策略崩塌。作者把课程学习搬到 RL 微调上,想打破这个瓶颈。

Huilin Deng,Ding Zou,Xinghao Zhao,Rui Ma,Yanming Guo,Yang Cao,Yu Kang
vlmrlhfcurriculum-learningAnthologyDBLP
6
泛读EMNLP 2025

Diverse, not Short: A Length-Controlled Data Selection Strategy for Improving Response Diversity of Language Models

SFT 后的模型回答千篇一律:现有'提升多样性'的数据选择策略经常只是偏向了长回答,把'长'混淆成了'多样'。作者想把长度因素剔除,看真正的多样性选择该怎么做。

Vijeta Deshpande,Debasmita Ghose,John D. Patterson,Roger E. Beaty,Anna Rumshisky
data-selectionlength-controldiversityDOIDBLP
6
泛读FindingsEMNLP 2025

QA-LIGN: Aligning LLMs through Constitutionally Decomposed QA

基于宪法 AI(constitutional AI)的对齐通常把原则作为整体喂给模型,粒度太粗,评估和奖励都很糊。作者把宪法原则拆成一组 QA,让对齐过程可分解、可验证。

Jacob Dineen,Aswin RRV,Qin Liu,Zhikun Xu,Xiao Ye,Ming Shen ... 省略 1 位作者 ... ,Shijie Lu,Chitta Baral,Muhao Chen,Ben Zhou
alignmentconstitutionqaAnthologyDBLP
6
泛读EMNLP 2025

SDGO: Self-Discrimination-Guided Optimization for Consistent Safety in Large Language Models

LLM 的安全性在不同输入形式下不一致:同一个有害意图换个表达方式(越狱 prompt、多语、改写)就能绕过。作者想让模型自己辨别这种一致性问题并自我优化。

Peng Ding,Wen Sun,Dailin Li,Wei Zou,Jiaming Wang,Jiajun Chen,Shujian Huang
safetyalignmentoptimizationDOIDBLP
6
泛读EMNLP 2025

From Problem-Solving to Teaching Problem-Solving: Aligning LLMs with Pedagogy using Reinforcement Learning

让 LLM 从“解题者”变成“教学者”——在学生求助时不要直接给答案,而要引导。之前用 SFT 模仿教师对话的做法泛化差,因为教学行为是多轮、长期目标,单步模仿很难学到“何时不给答案”。

David Dinucu-Jianu,Jakub Macina,Nico Daheim,Ido Hakimi,Iryna Gurevych,Mrinmaya Sachan
TU Darmstadt (UKP Lab)ETH Zürichreinforcement-learningalignmentpedagogyDOIDBLP
6
泛读EMNLP 2025

SPIRIT: Patching Speech Language Models against Jailbreak Attacks

语音 LM 的越狱防护还几乎是空白。文本 LLM 的对齐/防越狱研究很多,但当输入变成语音(音频 token 或 waveform),攻击面新增了韵律、噪声、说话风格等维度,已有文本侧防御直接失效。

Amirbek Djanibekov,Nurdaulet Mukhituly,Kentaro Inui,Hanan Aldarmaki,Nils Lukas
MBZUAIspeech-lmjailbreaksafetyDOIDBLP
6
泛读FindingsEMNLP 2025

Sample Efficient Alignment Learning With Episodic Control

RLHF/偏好对齐样本效率太低——每条偏好数据都要更新策略,大量相似 prompt 上的经验没被复用。

Van Dai Do,Quan Hung Tran,Ahmed Kirmani,Lu Zhang,Hung Le
Deakin Universityalignmentsample-efficiencyreinforcement-learningAnthologyDBLP
6
泛读EMNLP 2025

Beyond A Single AI Cluster: A Survey of Decentralized LLM Training

去中心化 LLM 训练正在从玩票走向严肃方案——单集群受限于能源、芯片获取和带宽成本,但跨地域训练有慢互联、异构硬件、容错等开放问题。缺一份把这些问题梳理清楚的 survey。

Haotian Dong,Jingyan Jiang,Rongwei Lu,Jiajun Luo,Jiajun Song,Bowen Li,Ying Shen,Zhi Wang
distributed-trainingdecentralizedsystemsDOIDBLP
6
泛读EMNLP 2025

Studying the Role of Input-Neighbor Overlap in Retrieval-Augmented Language Models Training Efficiency

RALM(检索增强 LM)训练时,检索回来的邻居和输入高度重叠——模型到底是学到了利用检索,还是只是用邻居当作答案的近似 copy?重叠度对训练效率和泛化的影响缺少系统研究。

Ehsan Doostmohammadi,Marco Kuhlmann
Linköping Universityragtraining-efficiencyretrievalDOIDBLP
6
泛读EMNLP 2025

Agentic-R1: Distilled Dual-Strategy Reasoning

推理型模型要么擅长长 CoT(慢思考、工具少),要么擅长 agentic 工具调用(动作多、推理浅),单个模型很难两者兼得。R1 类蒸馏常常只得到一种风格。

Weihua Du,Pranjal Aggarwal,Sean Welleck,Yiming Yang
CMUreasoningdistillationr1DOIDBLP
6
泛读EMNLP 2025

DVAGen: Dynamic Vocabulary Augmented Generation

固定词表限制了生成中的实体、术语、长尾符号表达——每次遇到 OOV 就被切成若干 subword,生成冗长且易错。copy 机制和 pointer network 解决了一部分但未统一到现代 LLM 解码框架。

Wei Du,Nuowei Liu,Jie Wang,Jiahao Kuang,Tao Ji,Xiaoling Wang,Yuanbin Wu
vocabularygenerationdecodingDOIDBLP
6
泛读EMNLP 2025

ReSURE: Regularizing Supervision Unreliability for Multi-turn Dialogue Fine-tuning

Yiming Du,Yifan Xiang,Bin Liang,Dahua Lin,Kam-Fai Wong,Fei Tan
dialoguefine-tuningdata-qualityDOIDBLP
6
泛读EMNLP 2025

CopySpec: Accelerating LLMs with Speculative Copy-and-Paste

Razvan-Gabriel Dumitru,Minglai Yang,Vikas Yadav,Mihai Surdeanu
speculative-decodingcopyinferenceDOIDBLP
6
泛读EMNLP 2025

Which Word Orders Facilitate Length Generalization in LMs? An Investigation with GCG-Based Artificial Languages

Nadine El-Naggar,Tatsuki Kuribayashi,Ted Briscoe
length-generalizationword-orderartificial-languagesDOIDBLP
6
泛读EMNLP 2025

Mixture-of-Clustered-Experts: Advancing Expert Specialization and Generalization in Instruction Tuning

Sugyeong Eo,Jung Jun Lee,Chanjun Park,Heuiseok Lim
moeinstruction-tuningexpert-specializationDOIDBLP
6
泛读FindingsEMNLP 2025

Layer Duplication in LLMs

预训练后扩大模型规模通常需要昂贵的从头训练,如何在不破坏已有表征能力的前提下,通过复用已有权重实现低成本的参数扩展?

Neo Eyal,Nachum Dershowitz,Kfir Bar
Tel Aviv Universitylayer-duplicationarchitectureAnthologyDBLP
6
泛读FindingsEMNLP 2025

Hopscotch: Discovering and Skipping Redundancies in Language Models

LLM 推理时所有 token 都必须经过所有 Transformer 层,但大量简单 token 的表征在浅层就已饱和,导致严重的计算冗余。

Mustafa Eyceoz,Nikhil Shivakumar Nayak,Hao Wang,Ligong Han,Akash Srivastava
MITIBMlayer-skippingredundancyinferenceAnthologyDBLP
6
泛读EMNLP 2025

The Medium Is Not the Message: Deconfounding Document Embeddings via Linear Concept Erasure

文档 Embedding 经常将语义内容与文档长度、作者风格、来源等混杂变量(Confounders)纠缠在一起,导致检索结果偏离真实语义需求。

Yu Fan,Yang Tian,Shauli Ravfogel,Mrinmaya Sachan,Elliott Ash,Alexander Miserlis Hoyle
ETH ZurichBar-Ilan UniversityUniversity of Washingtonembeddingsconcept-erasuredeconfoundingDOIDBLP
6
泛读EMNLP 2025

To See a World in a Spark of Neuron: Disentangling Multi-Task Interference for Training-Free Model Merging

在进行免训练的模型合并(Model Merging,如 Task Arithmetic)时,不同任务的权重更新会产生严重的参数干扰,导致合并后模型性能大打折扣。

Zitao Fang,Guodong Du,Shuyang Yu,Yifei Guo,Yiwei Zhang,Yiyao Cao,Jing Li,Ho-Kin Tang,Sim Kuan Goh
model-mergingneuron-analysismulti-taskDOIDBLP
6
泛读FindingsEMNLP 2025

Safeguard Fine-Tuned LLMs Through Pre- and Post-Tuning Model Merging

这篇工作要解决的是:安全对齐微调常常损伤通用能力,而继续追能力又容易把安全性冲掉,作者试图用微调前后模型合并来缓解这种张力。过去更常见的做法是直接在单一 checkpoint 上叠加安全数据或拒答数据,但这种方式容易产生灾难性遗忘或安全脆弱性。

Hua Farn,Hsuan Su,Shachi H. Kumar,Saurav Sahay,Shang-Tse Chen,Hung-yi Lee
model-mergingsafetyfine-tuningAnthologyDBLP
6
泛读FindingsEMNLP 2025

Pruning Weights but Not Truth: Safeguarding Truthfulness While Pruning LLMs

这篇工作要解决的是:LLM 剪枝常能保住常规 benchmark 分数,却容易损伤 truthfulness,尤其是在不确定知识和诱导性提问下更明显。作者要证明并缓解一个常被忽略的问题:压缩模型不只是掉能力,也可能更系统地掉‘说真话’的倾向。

Yao Fu,Runchao Li,Xianxuan Long,Haotian Yu,Xiaotian Han,Yu Yin,Pan Li
pruningtruthfulnessAnthologyDBLP
6
泛读EMNLP 2025

Quantized but Deceptive? A Multi-Dimensional Truthfulness Evaluation of Quantized LLMs

Yao Fu,Xianxuan Long,Runchao Li,Haotian Yu,Mu Sheng,Xiaotian Han,Yu Yin,Pan Li
quantizationtruthfulnessDOIDBLP
6
泛读EMNLP 2025

Sparse Autoencoder Features for Classifications and Transferability

Jack Gallifant,Shan Chen,Kuleen Sasse,Hugo J. W. L. Aerts,Thomas Hartvigsen,Danielle S. Bitterman
saeinterpretabilityfeaturesDOIDBLP
6
泛读FindingsEMNLP 2025

MASSV: Multimodal Adaptation and Self-Data Distillation for Speculative Decoding of Vision-Language Models

Mugilan Ganesan,Shane Segal,Ankur Aggarwal,Nish Sinnadurai,Sean Lie,Vithursan Thangarasa
speculative-decodingvlmdistillationAnthologyDBLP
6
泛读EMNLP 2025

Group-SAE: Efficient Training of Sparse Autoencoders for Large Language Models via Layer Groups

Davide Ghilardi,Federico Belotti,Marco Molinari,Tao Ma,Matteo Palmonari
sparse-autoencoderinterpretabilityefficient-trainingDOIDBLP
6
泛读FindingsEMNLP 2025

Domain Pre-training Impact on Representations

César González-Gutiérrez,Ariadna Quattoni
domain-pretrainingrepresentationcontinual-pretrainAnthologyDBLP
6
泛读FindingsEMNLP 2025

Token Preference Optimization with Self-Calibrated Visual-Anchored Rewards for Hallucination Mitigation

Jihao Gu,Yingyao Wang,Meng Cao,Pi Bu,Jun Song,Bo Zheng,Yancheng He,Shilong Li
dpovlmhallucinationAnthologyDBLP
6
泛读EMNLP 2025

From General Reward to Targeted Reward: Improving Open-ended Long-context Generation Models

解决长文本生成任务中,全局单一的 Reward 信号过于稀疏,导致模型难以获得细粒度优化方向的问题。

Zhihan Guo,Jiele Wu,Wenqian Cui,Yifei Zhang,Minda Hu,Yufei Wang,Irwin King
reward-modellong-contextgenerationDOIDBLP
6
泛读EMNLP 2025

CROP: Contextual Region-Oriented Visual Token Pruning

解决视觉 Token 剪枝(Token Pruning)过程中,容易破坏图像局部区域上下文(Region-Oriented Context)和空间结构的问题。

Jiawei Guo,Feifei Zhai,Pu Jian,Qianrun Wei,Yu Zhou
visual-tokenizertoken-pruningvlmDOIDBLP
6
泛读EMNLP 2025

DSG-MCTS: A Dynamic Strategy-Guided Monte Carlo Tree Search for Diversified Reasoning in Large Language Models

解决 LLM 在复杂推理任务中使用传统 MCTS(蒙特卡洛树搜索)时,探索策略单一导致容易陷入局部最优或搜索效率低下的问题。

Rui Ha,Chaozhuo Li,Rui Pu,Litian Zhang,Xi Zhang,Sen Su
mctsreasoninginference-scalingDOIDBLP
6
泛读FindingsEMNLP 2025

Identifying Noise in Human-Created Datasets using Training Dynamics from Generative Models

解决人工标注数据集中存在大量隐性噪声,且难以通过静态规则进行大规模自动化清洗的问题。

Maeda F. Hanafi,Ishan Jindal,Yannis Katsis,Lucian Popa,Huaiyu Zhu
IBM Researchdata-qualitytraining-dynamicsnoiseAnthologyDBLP
6
泛读EMNLP 2025

TinySQL: A Progressive Text-to-SQL Dataset for Mechanistic Interpretability Research

这篇论文要解决的是:机制可解释性研究缺少一个足够可控、可分阶段、又带真实结构约束的 text-to-SQL 数据集。现有数据集要么过于真实导致变量太多,要么过于玩具化无法承载电路级分析,因此很难定位模型到底学会了哪一步。

Abir Harrasse,Philip Quirke,Clement Neo,Dhruv Nathawani,Luke Marks,Amir Abdullah
mechanistic-interpretabilitydatasetsqlDOIDBLP
6
泛读EMNLP 2025

Router-Tuning: A Simple and Effective Approach for Dynamic Depth

这篇论文关注动态深度的一个现实障碍:怎样在不重训整网、不过多改动架构的情况下,让模型按输入难度自适应选择计算量。已有动态深度方法往往引入额外控制模块、训练复杂,或者收益被路由误差吃掉。

Shwai He,Tao Ge,Guoheng Sun,Bowei Tian,Xiaoyang Wang,Dong Yu
dynamic-depthroutingefficiencyDOIDBLP
6
泛读EMNLP 2025

Order Doesn't Matter, But Reasoning Does: Training LLMs with Order-Centric Augmentation

这篇论文讨论的是:推理训练里,表面上的步骤顺序变化可能不重要,但推理结构本身重要;因此能否用 order-centric augmentation 提升模型对推理本质的把握。很多 CoT 训练默认固定推理顺序,结果模型可能把顺序模式当捷径记住,而不是学会真正的依赖关系。

Qianxi He,Qianyu He,Jiaqing Liang,Weikang Zhou,Zeye Sun,Fei Yu,Yanghua Xiao
reasoningdata-augmentationorderDOIDBLP
6
泛读EMNLP 2025

SAE-SSV: Supervised Steering in Sparse Representation Spaces for Reliable Control of Language Models

这篇论文想解决的是:怎样更可靠地控制语言模型行为,同时避免直接在 dense hidden state 上做 steering 带来的副作用和不稳定。已有激活编辑或向量 steering 往往有效但脆弱,容易牺牲无关能力,且控制边界不清楚。

Zirui He,Mingyu Jin,Bo Shen,Ali Payani,Yongfeng Zhang,Mengnan Du
saesteeringrepresentationDOIDBLP
6
泛读FindingsEMNLP 2025

Efficiently Editing Mixture-of-Experts Models with Compressed Experts

这篇论文要解决的是:MoE 模型编辑成本高、定位难,而每个 expert 又天然带有参数冗余,能否在压缩表示下高效编辑。传统模型编辑大多围绕 dense 模型设计,直接迁移到 MoE 往往要面对 expert 选择、参数分散和编辑局部性差的问题。

Yifei He,Yang Liu,Chen Liang,Hany Hassan Awadalla
moemodel-editingcompressionAnthologyDBLP
6
泛读EMNLP 2025

Seeing Through Words, Speaking Through Pixels: Deep Representational Alignment Between Vision and Language Models

这篇工作要回答的是:视觉模型和语言模型内部表征到底有没有深层对齐,而不是只在最终任务头上表现出表面可迁移性。以往很多工作主要看跨模态检索、zero-shot 分类或线性 probe,能说明输出层可对齐,但很难判断两种模型是否在中间层学到了相近的概念结构,因此这个问题值得单独拆开分析。

Zoe Wanying He,Sean Trott,Meenakshi Khosla
representationalignmentvision-languageDOIDBLP
6
泛读FindingsEMNLP 2025

RingFormer: Rethinking Recurrent Transformer with Adaptive Level Signals

这篇工作讨论的是:如何重新设计 recurrent transformer,使它在保留递归状态效率的同时减少长程建模中的信息退化。过去 recurrent transformer 的主要问题是状态压缩过强、层间信号单一,导致它虽然省算省缓存,但在复杂上下文上容易丢层级结构。

Jaemu Heo,Eldor Fozilov,Hyunmin Song,Taehwan Kim
transformerrecurrentarchitectureAnthologyDBLP
6
泛读FindingsEMNLP 2025

Spelling-out is not Straightforward: LLMs' Capability of Tokenization from Token to Characters

这篇工作要回答的是:LLM 是否真的理解 token 到字符的映射,还是只在常见词上记住了表面模式。这个问题以前常被低估,因为很多评测默认 tokenizer 是透明的,而实际中 tokenization 会显著影响拼写、计数、反向字符串和代码等任务表现。

Tatsuya Hiraoka,Kentaro Inui
tokenizercharacter-levelanalysisAnthologyDBLP
6
泛读EMNLP 2025

Improving Neutral Point-of-View Generation with Data- and Parameter-Efficient RL

这篇工作要解决的是:如何用更省数据、也更省参数更新的方式,提高模型生成中立观点表述的能力。传统做法通常依赖大规模偏好数据和全参数 RLHF,但在 neutral POV 这种相对窄但高敏感的目标上,这样做成本高、容易过拟合,还可能顺带伤害一般能力。

Jessica Hoffmann,Christiane Ahlheim,Zac Yu,Aria Walfrand,Jarvis Jin,Marie Tano ... 省略 1 位作者 ... ,Erin MacMurray van Liemt,Nithum Thain,Hakim Sidahmed,Lucas Dixon
rlalignmentpeftDOIDBLP
6
泛读EMNLP 2025

Slim-SC: Thought Pruning for Efficient Scaling with Self-Consistency

这篇工作要解决的是:self-consistency 通过采样多条思维链能提高推理准确率,但推理成本很高,很多 sampled thoughts 实际上是冗余的,能否在不明显掉点的情况下做 thought pruning。过去的优化多集中在减少样本数,但这会粗暴牺牲多样性;更细粒度的思路是保留有用思路、剪掉无效思路。

Colin Hong,Xu Guo,Anand Chaanan Singh,Esha Choukse,Dmitrii Ustiugov
self-consistencyinferencereasoningDOIDBLP
6
泛读EMNLP 2025

Intrinsic Test of Unlearning Using Parametric Knowledge Traces

Yihuai Hong,Lei Yu,Haiqin Yang,Shauli Ravfogel,Mor Geva
unlearningknowledgeinterpretabilityDOIDBLP
6
泛读EMNLP 2025

Evaluating Robustness of Large Audio Language Models to Audio Injection: An Empirical Study

Guanyu Hou,Jiaming He,Yinhang Zhou,Ji Guo,Yitong Qiao,Rui Zhang,Wenbo Jiang
audio-lmrobustnesssafetyDOIDBLP
6
泛读FindingsEMNLP 2025

LNE-Blocking: An Efficient Framework for Contamination Mitigation Evaluation on Large Language Models

Ruijie Hou,Jiao Yueyang,Hanxu Hu,Yingming Li,Wai Lam,Huajian Zhang,Hongyuan Lu
contaminationevaluationbenchmarkAnthologyDBLP
6
泛读EMNLP 2025

MLWQ: Efficient Small Language Model Deployment via Multi-Level Weight Quantization

Chun Hu,Junhui He,Shangyu Wu,Yuxin He,Chun Jason Xue,Qingan Li
quantizationsmall-modeldeploymentDOIDBLP
6
泛读FindingsEMNLP 2025

Steering LVLMs via Sparse Autoencoder for Hallucination Mitigation

Zhenglin Hua,Jinghan He,Zijun Yao,Tianxu Han,Haiyun Guo,Yuheng Jia,Junfeng Fang
saehallucinationvlmAnthologyDBLP
6
泛读EMNLP 2025

Flaw or Artifact? Rethinking Prompt Sensitivity in Evaluating LLMs

Andong Hua,Kenan Tang,Chenhe Gu,Jindong Gu,Eric Wong,Yao Qin
evaluationprompt-sensitivityDOIDBLP
6
泛读EMNLP 2025

Mitigating Catastrophic Forgetting in Large Language Models with Forgetting-aware Pruning

Wei Huang,Anda Cheng,Yinggui Wang
pruningcatastrophic-forgettingcontinualDOIDBLP
6
泛读FindingsEMNLP 2025

DCRM: A Heuristic to Measure Response Pair Quality in Preference Optimization

Chengyu Huang,Tanya Goyal
dpopreferencedata-qualityAnthologyDBLP
6
泛读EMNLP 2025

Governance in Motion: Co-evolution of Constitutions and AI models for Scalable Safety

Chenhao Huang,Ziyu Shen,Yicong Ren,Huiyuan Zheng,Jiazheng Zhang,Mingxu Chai ... 省略 3 位作者 ... ,Jie Shi,Tao Gui,Qi Zhang,Xuanjing Huang
safetyalignmentco-evolutionDOIDBLP
6
泛读EMNLP 2025

Path Drift in Large Reasoning Models: How First-Person Commitments Override Safety

Yuyi Huang,Runzhe Zhan,Lidia S. Chao,Ailin Tao,Derek F. Wong
safetyreasoningjailbreakDOIDBLP
6
泛读FindingsEMNLP 2025

SelfAug: Mitigating Catastrophic Forgetting in Retrieval-Augmented Generation via Distribution Self-Alignment

Yuqing Huang,Rongyang Zhang,Qimeng Wang,Chengqiang Lu,Yan Gao,Yi Wu ... 省略 2 位作者 ... ,Guiquan Liu,Xin Li,Hao Wang,Enhong Chen
catastrophic-forgettingragcontinualAnthologyDBLP
6
泛读EMNLP 2025

Improving Instruct Models for Free: A Study on Partial Adaptation

Ozan Irsoy,Pengxiang Cheng,Jennifer L. Chen,Daniel Preotiuc-Pietro,Shiyue Zhang,Duccio Pappadopulo
partial-adaptationsftinstruct-tuningDOIDBLP
6
泛读FindingsEMNLP 2025

FiRST: Finetuning Router-Selective Transformers for Input-Adaptive Latency Reduction

Akriti Jain,Saransh Sharma,Koyel Mukherjee,Soumyabrata Pal
early-exitadaptive-inferencerouterAnthologyDBLP
6
泛读EMNLP 2025

Balcony: A Lightweight Approach to Dynamic Inference of Generative Language Models

Benyamin Jamialahmadi,Parsa Kavehzadeh,Mehdi Rezagholizadeh,Parsa Farinneya,Hossein Rajabzadeh,Aref Jafari,Boxing Chen,Marzieh S. Tahaei
dynamic-inferenceearly-exitefficiencyDOIDBLP
6
泛读EMNLP 2025

Mixture of Languages: Improved Multilingual Encoders Through Language Grouping

解决多语言预训练中不同语系之间因共享参数导致的容量瓶颈和相互干扰(多语言诅咒)问题。

João Maria Janeiro,Belen Alastruey,Francisco Massa,Maha Elbayad,Benjamin Piwowarski,Patrick Gallinari,Loïc Barrault
Meta AIKyutaiSorbonne Universitémultilingualencoderlanguage-groupingDOIDBLP
6
泛读FindingsEMNLP 2025

Self-Training Large Language Models with Confident Reasoning

解决大模型自我训练(Self-Training)过程中,因采纳错误推理路径(即使最终答案正确)而导致的确认偏差(Confirmation Bias)。

Hyosoon Jang,Yunhui Jang,Sungjae Lee,Jungseul Ok,Sungsoo Ahn
POSTECHSeoul National Universityself-trainingreasoningconfidenceAnthologyDBLP
6
泛读FindingsEMNLP 2025

Public Data Assisted Differentially Private In-Context Learning

在保护演示数据隐私的前提下,如何提升差分隐私上下文学习(DP-ICL)的效用。

Seongho Joo,Hyukhun Koh,Kyomin Jung
Seoul National Universityin-context-learningdifferential-privacydata-qualityAnthologyDBLP
6
泛读EMNLP 2025

Calibration Across Layers: Understanding Calibration Evolution in LLMs

分析大模型在逐层前向传播过程中,其输出置信度(Calibration)是如何演变的。

Abhinav Joshi,Areeb Ahmad,Ashutosh Modi
IIT Kanpurcalibrationlayer-analysistraining-dynamicsDOIDBLP
6
泛读FindingsEMNLP 2025

From n-gram to Attention: How Model Architectures Learn and Propagate Bias in Language Modeling

这篇工作要回答的是:不同语言模型架构究竟如何学习并传播语料偏差,而不仅仅是“是否有偏差”。以往偏差研究常把模型当黑盒测输出差异,较少追问 n-gram、RNN、Transformer/attention 在偏差吸收和传播路径上的机制差异;作者显然在做这种架构层面的拆解。

Mohsinul Kabir,Tasfia Tahsin,Sophia Ananiadou
biaslanguage-modelingarchitectureAnthologyDBLP
6
泛读EMNLP 2025

Pre-trained Language Models Learn Remarkably Accurate Representations of Numbers

这篇论文要回答的是:预训练语言模型到底有没有学到准确的数值表示,而不只是记住常见数字模式。过去对 LLM 数字能力的判断通常来自算术 benchmark,结论多半偏悲观,因为模型在计算任务上错误很多;作者把问题收窄到“表示是否准确”,并给出一个相反方向的判断:预训练模型可能比外界以为的更会表示数字。

Marek Kadlcík,Michal Stefánik,Timothee Mickus,Josef Kuchar,Michal Spiegel
representationnumberspretrained-modelsDOIDBLP
6
泛读EMNLP 2025

Balanced Multi-Factor In-Context Learning for Multilingual Large Language Models

这篇工作要解决的是:多语大模型做 in-context learning 时,示例选择往往在多个因素上失衡,导致性能不稳定。现有 ICL 通常只按语义相似度、语言一致性或标签覆盖单独选例,结果是某一因素优化了,另一因素被破坏;作者提出 balanced multi-factor ICL 来同时平衡这些条件。

Masahiro Kaneko,Alham Fikri Aji,Timothy Baldwin
in-context-learningmultilingualpromptingDOIDBLP
6
泛读EMNLP 2025

Can GRPO Boost Complex Multimodal Table Understanding?

Xiaoqiang Kang,Shengen Wu,Zimu Wang,Yilin Liu,Xiaobo Jin,Kaizhu Huang,Wei Wang,Yutao Yue,Xiaowei Huang,Qiufeng Wang
grpomultimodalreasoningDOIDBLP
6
泛读EMNLP 2025

REVIVING YOUR MNEME: Predicting The Side Effects of LLM Unlearning and Fine-Tuning via Sparse Model Diffing

Aly M. Kassem,Zhuan Shi,Negar Rostamzadeh,Golnoosh Farnadi
unlearningfine-tuningmodel-diffingDOIDBLP
6
泛读EMNLP 2025

AQuilt: Weaving Logic and Self-Inspection into Low-Cost, High-Relevance Data Synthesis for Specialist LLMs

Xiaopeng Ke,Hexuan Deng,Xuebo Liu,Jun Rao,Zhenxi Song,Jun Yu,Min Zhang
data-synthesissftspecialistDOIDBLP
6
泛读EMNLP 2025

Demystifying Domain-adaptive Post-training for Financial LLMs

Zixuan Ke,Yifei Ming,Xuan-Phi Nguyen,Caiming Xiong,Shafiq Joty
continual-pretraindomain-adaptationfinanceDOIDBLP
6
泛读EMNLP 2025

When Life Gives You Samples: The Benefits of Scaling up Inference Compute for Multilingual LLMs

Ammar Khairi,Daniel D'souza,Ye Shen,Julia Kreutzer,Sara Hooker
inference-scalingmultilingualtest-timeDOIDBLP
6
泛读IndustryEMNLP 2025

DACIP-RC: Domain Adaptive Continual Instruction Pre-Training via Reading Comprehension on Business Conversations

Elena Khasanova,Harsh Saini,Md. Tahmid Rahman Laskar,Xue-Yong Fu,Cheng Chen,Shashi Bhushan TN
continual-pretraindomain-adaptationinstructionDOIDBLP
6
泛读IndustryEMNLP 2025

ixi-GEN: Efficient Industrial sLLMs through Domain Adaptive Continual Pretraining

Seonwu Kim,Yohan Na,Kihun Kim,Hanhee Cho,Geun Lim,Mintae Kim,Seongik Park,Ki Hyun Kim,Youngsub Han,Byoung-Ki Jeon
continual-pretraindomain-adaptationsmall-lmDOIDBLP
6
泛读EMNLP 2025

Benchmark Profiling: Mechanistic Diagnosis of LLM Benchmarks

Dongjun Kim,Gyuho Shim,Yongchan Chun,Minhyuk Kim,Chanjun Park,Heuiseok Lim
benchmarkevaluationdiagnosisDOIDBLP
6
泛读FindingsEMNLP 2025

Leveraging High-Resource English Corpora for Cross-lingual Domain Adaptation in Low-Resource Japanese Medicine via Continued Pre-training

Kazuma Kobayashi,Zhen Wan,Fei Cheng,Tsuta Yuma,Xin Zhao,Junfeng Jiang ... 省略 4 位作者 ... ,Yuki Arase,Daisuke Kawahara,Akiko Aizawa,Sadao Kurohashi
continual-pretrainmultilingualdomain-adaptationAnthologyDBLP
6
泛读EMNLP 2025

Learning from Diverse Reasoning Paths with Routing and Collaboration

现有的推理模型微调通常只学习单一的“黄金路径”,忽略了多样化推理路径的价值,导致模型在遇到分布外问题时缺乏鲁棒性和灵活性。

Zhenyu Lei,Zhen Tan,Song Wang,Yaochen Zhu,Zihan Chen,Yushun Dong,Jundong Li
reasoningroutingcollaborationDOIDBLP
6
泛读EMNLP 2025

Analysing Chain of Thought Dynamics: Active Guidance or Unfaithful Post-hoc Rationalisation?

CoT 究竟是真正在引导模型一步步得出结论(Active Guidance),还是模型在隐层已经得出了结论,CoT 只是在做马后炮式的合理化解释(Post-hoc Rationalisation)?

Samuel Lewis-Lim,Xingwei Tan,Zhixue Zhao,Nikolaos Aletras
University of Sheffieldcotinterpretabilityreasoning-dynamicsDOIDBLP
6
泛读IndustryEMNLP 2025

L4: Mutual Learning Helps Lifelong Language Learning

这篇论文要解决的是持续语言学习里的遗忘与低效复用问题。现有 lifelong language learning 往往用单模型顺序训练或回放来缓解遗忘,但不同阶段学到的知识很难双向传递,结果是新任务学得快时旧能力掉得也快。

Jiyong Li,Dilshod Azizov,Shangsong Liang
continual-learninglifelong-learningDOIDBLP
6
泛读FindingsEMNLP 2025

Language Models Can Easily Learn to Reason from Demonstrations

这篇论文想回答一个很直接但有争议的问题:语言模型的“推理”能力,到底需要专门训练出来,还是从 demonstrations 就能很容易学会。此前大量工作把 reasoning 提升归因于链式思维数据、强化学习或专门合成语料,但这可能混淆了能力习得和格式模仿。

Dacheng Li,Shiyi Cao,Tyler Griggs,Shu Liu,Xiangxi Mo,Eric Tang ... 省略 2 位作者 ... ,Shishir G. Patil,Matei Zaharia,Joseph E. Gonzalez,Ion Stoica
in-context-learningreasoningdemonstrationsAnthologyDBLP
6
泛读EMNLP 2025

Diagnosing Memorization in Chain-of-Thought Reasoning, One Token at a Time

这篇论文要解决的是:CoT 推理里到底有多少是逐步计算,多少是逐 token 记忆复现。此前关于 CoT memorization 的讨论多停留在样例级或答案级,难以定位模型是在推理链的哪一段开始依赖训练记忆。

Huihan Li,You Chen,Siyuan Wang,Yixin He,Ninareh Mehrabi,Rahul Gupta,Xiang Ren
memorizationcottoken-level-analysisDOIDBLP
6
泛读EMNLP 2025

Add-One-In: Incremental Sample Selection for Large Language Models via a Choice-Based Greedy Paradigm

这篇论文要解决的是大模型训练样本选择的增量优化问题:在不能穷举所有子集的前提下,怎样一步一步加入更有价值的样本。过去常见做法是打分排序或一次性子集选择,但它们难以建模样本间互补性,容易选到一堆彼此重复的“高分”样本。

Zhuo Li,Yuhao Du,Xiaoqi Jiao,Steven Y. Guo,Yuege Feng,Xiang Wan,Anningzhe Gao,Jinpeng Hu
data-selectionincremental-learningsftDOIDBLP
6
泛读FindingsEMNLP 2025

FAEDKV: Infinite-Window Fourier Transform for Unbiased KV Cache Compression

这篇论文要解决的是长上下文推理里的 KV cache 压缩偏差问题。现有压缩方法大多通过窗口截断、聚类或低秩近似节省显存,但一旦只保留有限窗口或做有偏摘要,模型对远程依赖的注意力会系统性失真。

Runchao Li,Yao Fu,Mu Sheng,Xianxuan Long,Haotian Yu,Pan Li
kv-cachecompressionlong-contextAnthologyDBLP
6
泛读FindingsEMNLP 2025

LoRaDA: Low-Rank Direct Attention Adaptation for Efficient LLM Fine-tuning

这篇论文要解决的是 LLM 微调里低秩适配对注意力模块改动不够直接、效率与效果之间不够理想的问题。传统 LoRA 通常挂在线性投影上,虽然便宜,但它只能间接影响 attention pattern,某些任务下表达力不够。

Zhangming Li,Qinghao Hu,Yiqun Chen,Peisong Wang,Yifan Zhang,Jian Cheng
peftloraattentionAnthologyDBLP
6
泛读IndustryEMNLP 2025

Encouraging Good Processes Without the Need for Good Answers: Reinforcement Learning for LLM Agent Planning

这篇论文要解决的是 agent planning 的 RL 训练中,为什么非得依赖“好答案”才能学到好过程。很多规划任务很难拿到最终高质量答案或环境回报稀疏,结果 RL 训练信号要么太贵,要么几乎传不过去。

Zhiwei Li,Yong Hu,Wenqing Wang
reinforcement-learningprocess-rewardplanningDOIDBLP
6
泛读EMNLP 2025

GraphKV: Breaking the Static Selection Paradigm with Graph-Based KV Cache Eviction

Xuelin Li,Xiangqi Jin,Linfeng Zhang
kv-cacheevictionefficiencyDOIDBLP
6
泛读FindingsEMNLP 2025

Teaching According to Talents! Instruction Tuning LLMs with Competence-Aware Curriculum Learning

Yangning Li,Tingwei Lu,Yinghui Li,Yankai Chen,Wei-Chieh Huang,Wenhao Jiang,Hui Wang,Hai-Tao Zheng,Philip S. Yu
instruction-tuningcurriculum-learningsftAnthologyDBLP
6
泛读FindingsEMNLP 2025

R3-RAG: Learning Step-by-Step Reasoning and Retrieval for LLMs via Reinforcement Learning

Yuan Li,Qi Luo,Xiaonan Li,Bufan Li,Qinyuan Cheng,Bo Wang,Yining Zheng,Yuxin Wang,Zhangyue Yin,Xipeng Qiu
reinforcement-learningragreasoningAnthologyDBLP
6
泛读EMNLP 2025

Layer-Aware Representation Filtering: Purifying Finetuning Data to Preserve LLM Safety Alignment

Hao Li,Lijun Li,Zhenghao Lu,Xianyi Wei,Rui Li,Jing Shao,Lei Sha
data-filteringsafety-alignmentrepresentationDOIDBLP
6
泛读FindingsEMNLP 2025

Text or Pixels? Evaluating Efficiency and Understanding of LLMs with Visual Text Inputs

Yanhong Li,Zixuan Lan,Jiawei Zhou
multimodalvisual-textefficiencyAnthologyDBLP
6
泛读FindingsEMNLP 2025

Treble Counterfactual VLMs: A Causal Approach to Hallucination

VLM 幻觉的归因一直模糊:到底是视觉通路看错了、语言先验压倒了视觉,还是两者交互出了问题?已有方法多靠相关性诊断,因果关系讲不清。

Li Li,Jiashu Qu,Linxin Song,Yuxiao Zhou,Yuehan Qin,Tiankai Yang,Yue Zhao
vlmhallucinationcausalAnthologyDBLP
6
泛读FindingsEMNLP 2025

Speculative Decoding for Multi-Sample Inference

推测解码(speculative decoding)目前都是为单样本生成设计的,但推理时常需要 multi-sample(比如 self-consistency、best-of-N)。在这种场景下重复小模型起草和大模型验证的逻辑没有被复用,开销被放大。

Yiwei Li,Jiayi Shi,Shaoxiong Feng,Peiwen Yuan,Xinglin Wang,Yueqi Zhang ... 省略 1 位作者 ... ,Chuyi Tan,Boyuan Pan,Yao Hu,Kan Li
speculative-decodinginferencesamplingAnthologyDBLP
6
泛读FindingsEMNLP 2025

LongTableBench: Benchmarking Long-Context Table Reasoning across Real-World Formats and Domains

现有表格推理 benchmark 基本都是短表、单一格式、学术域,和真实业务里的长表(多 sheet、财报、日志)脱节,导致 long-context 表格能力评估不可靠。

Liyao Li,Jiaming Tian,Hao Chen,Wentao Ye,Chao Ye,Haobo Wang,Ningtao Wang,Xing Fu,Gang Chen,Junbo Zhao
long-contextbenchmarktable-reasoningAnthologyDBLP
6
泛读FindingsEMNLP 2025

Lost in Embeddings: Information Loss in Vision-Language Models

VLM 把图像塞进 LLM 时,视觉 encoder 输出经过投影/重采样进入文本空间的过程中到底丢了多少信息、丢在哪,业界描述模糊。

Wenyan Li,Raphael Tang,Chengzu Li,Caiqi Zhang,Ivan Vulic,Anders Søgaard
University of Cambridgevlmembeddingsinformation-lossAnthologyDBLP
6
泛读EMNLP 2025

Reward-Shifted Speculative Sampling Is An Efficient Test-Time Weak-to-Strong Aligner

测试时对齐一般靠大模型重采样或 reward-guided decoding,成本高。能不能用小的 aligned 模型去引导大的 unaligned 模型,既省钱又对齐?

Bolian Li,Yanran Wu,Xinyu Luo,Ruqi Zhang
Purdue Universityspeculative-decodingsamplingalignmentDOIDBLP
6
泛读EMNLP 2025

Feature Extraction and Steering for Enhanced Chain-of-Thought Reasoning in Language Models

Zihao Li,Xu Wang,Yuzhe Yang,Ziyu Yao,Haoyi Xiong,Mengnan Du
cotsteeringfeature-analysisDOIDBLP
6
泛读EMNLP 2025

The Impact of Language Mixing on Bilingual LLM Reasoning

Yihao Li,Jiayi Xin,Miranda Muqing Miao,Qi Long,Lyle H. Ungar
multilingualreasoninglanguage-mixingDOIDBLP
6
泛读EMNLP 2025

LoRACoE: Improving Large Language Model via Composition-based LoRA Expert

Guanyu Li,Zhiheng Xi,Zhihao Zhang,Boyang Hong,Tao Gui,Qi Zhang,Xuanjing Huang
loramoeparameter-efficientDOIDBLP
6
泛读EMNLP 2025

START: Self-taught Reasoner with Tools

Chengpeng Li,Mingfeng Xue,Zhenru Zhang,Jiaxi Yang,Beichen Zhang,Bowen Yu,Binyuan Hui,Junyang Lin,Xiang Wang,Dayiheng Liu
tool-useself-trainingreasoningDOIDBLP
6
泛读EMNLP 2025

Towards Statistical Factuality Guarantee for Large Vision-Language Models

Zhuohang Li,Chao Yan,Nicholas J. Jackson,Wendi Cui,Bo Li,Jiaxin Zhang,Bradley A. Malin
vlmfactualityguaranteesDOIDBLP
6
泛读EMNLP 2025

Logits-Based Finetuning

Jingyao Li,Senqiao Yang,Sitong Wu,Han Shi,Chuanyang Zheng,Hong Xu,Jiaya Jia
distillationlogitsfinetuningDOIDBLP
6
泛读EMNLP 2025

Ambiguity Awareness Optimization: Towards Semantic Disambiguation for Direct Preference Optimization

DPO 默认所有偏好数据都是确定且二元的(chosen 绝对优于 rejected),忽略了人类偏好数据中固有的语义模糊性和噪声,导致模型容易过拟合噪声信号并降低生成多样性。

Jian Li,Shenglin Yin,Yujia Zhang,Alan Zhao,Xi Chen,Xiaohui Zhou,Pengfei Xu
dpopreference-optimizationDOIDBLP
6
泛读EMNLP 2025

Understanding the Thinking Process of Reasoning Models: A Perspective from Schoenfeld's Episode Theory

当前对推理模型(如 o1)的评估主要依赖最终答案的准确率,缺乏对模型生成长思维链(CoT)内部认知过程(如规划、验证、纠错)的结构化分析和理解。

Ming Li,Nan Zhang,Chenrui Fan,Hong Jiao,Yanbin Fu,Sydney Peters,Qingshu Xu,Robert Lissitz,Tianyi Zhou
reasoningcotanalysisDOIDBLP
6
泛读EMNLP 2025

A Training-Free Length Extrapolation Approach for LLMs: Greedy Attention Logit Interpolation

现有的免训练长度外推方法(如动态 RoPE 插值)通常会破坏局部注意力模式,导致模型在处理长文本时出现困惑度激增或局部信息丢失。

Yan Li,Tianyi Zhang,Zechuan Li,Caren Han
long-contextlength-extrapolationattentionDOIDBLP
6
泛读EMNLP 2025

Towards Infinite-Long Prefix in Transformer

在 Agent 或 RAG 场景中,处理极长甚至无限长的 System Prompt(前缀)会耗尽 KV Cache 内存,并导致极高的首字延迟(TTFT)。

Yingyu Liang,Zhenmei Shi,Zhao Song,Chiwun Yang
prefixlong-contexttransformerDOIDBLP
6
泛读FindingsEMNLP 2025

ThinkSwitcher: When to Think Hard, When to Think Fast

当前模型要么全部使用 System 1(直接生成,容易出错),要么全部使用 System 2(如 CoT 或 o1,在简单问题上浪费大量算力),缺乏动态分配计算资源的机制。

Guosheng Liang,Longguang Zhong,Ziyi Yang,Xiaojun Quan
reasoningadaptive-computecotAnthologyDBLP
6
泛读EMNLP 2025

E2LLM: Encoder Elongated Large Language Models for Long-Context Understanding and Reasoning

纯 Decoder 架构在处理超长上下文时受限于 O(N^2) 的注意力复杂度和庞大的 KV Cache,难以高效进行全局理解和推理。

Zihan Liao,Jun Wang,Hang Yu,Lingxiao Wei,Jianguo Li,Jun Wang,Wei Zhang
long-contextencoderarchitectureDOIDBLP
6
泛读FindingsEMNLP 2025

Reinforcement Learning with Supervised Alignment

纯粹的 RLHF 训练容易导致模型偏离预训练或 SFT 阶段的数据分布(即对齐税 Alignment Tax),表现为在通用能力(如 MMLU)上的性能下降。

João Luís Lins,Jia Xu
rlalignmentsftAnthologyDBLP
6
泛读EMNLP 2025

Information Integration in Large Language Models is Gated by Linguistic Structural Markers

这篇论文要回答的是:LLM 在整合句中信息时,并不是连续平滑地累加证据,而是会被语言结构标记显著门控。过去很多工作把信息整合看成更通用的注意力或记忆容量问题,这篇把焦点放到逗号、连接词、从句边界等结构提示上。

Wei Liu,Nai Ding
interpretabilitylinguisticsintegrationDOIDBLP
6
泛读EMNLP 2025

Mitigating Hallucinations in LM-Based TTS Models via Distribution Alignment Using GFlowNets

这篇论文解决的是 LM-based TTS 中的幻觉问题:模型会生成文本未给出的词、重复片段或不该出现的声学内容,而传统做法多靠解码约束或事后过滤,治标不治本。作者把问题重新表述为训练分布与目标语音-文本对齐分布不一致。

Chenlin Liu,Minghui Fang,Patrick Zhang,Wei Zhou,Jie Gao,Jiqing Han
ttsspeech-lmhallucinationDOIDBLP
6
泛读EMNLP 2025

VRoPE: Rotary Position Embedding for Video Large Language Models

这篇论文解决的是:标准 RoPE 为一维文本序列设计,直接搬到视频 LLM 时很难同时编码时间顺序与空间结构,导致长视频理解和帧间对齐受限。作者提出面向视频的 rotary position embedding 变体 VRoPE。

Zikang Liu,Longteng Guo,Yepeng Tang,Tongtian Yue,Junxian Cai,Kai Ma,Qingbin Liu,Xi Chen,Jing Liu
ropevideo-llmposition-embeddingDOIDBLP
6
泛读EMNLP 2025

AIR: Complex Instruction Generation via Automatic Iterative Refinement

这篇论文要解决的是复杂指令数据稀缺且人工编写成本高的问题,尤其是多约束、多步骤、易歧义的 instruction 数据难以规模化生产。现有自动合成方法往往一次生成就结束,复杂度和可执行性都不稳定。

Wei Liu,Yancheng He,Yu Li,Hui Huang,Chengwei Hu,Jiaheng Liu,Shilong Li,Wenbo Su,Bo Zheng
instruction-datasynthetic-dataDOIDBLP
6
泛读EMNLP 2025

CompassVerifier: A Unified and Robust Verifier for LLMs Evaluation and Outcome Reward

这篇论文要解决的是:LLM 评测 verifier 和 outcome reward model 往往各做各的,鲁棒性不足、任务迁移差,而且容易被表面格式或风格噪声带偏。作者希望构建一个统一且更稳的 verifier,同一套模型同时服务评测与奖励。

Shudong Liu,Hongwei Liu,Junnan Liu,Linchen Xiao,Songyang Gao,Chengqi Lyu ... 省略 1 位作者 ... ,Wenwei Zhang,Derek F. Wong,Songyang Zhang,Kai Chen
verifierreward-modelevaluationDOIDBLP
6
泛读EMNLP 2025

When Truthful Representations Flip Under Deceptive Instructions?

Xianxuan Long,Yao Fu,Runchao Li,Mu Sheng,Haotian Yu,Xiaotian Han,Pan Li
interpretabilitytruthfulnessalignmentDOIDBLP
6
泛读EMNLP 2025

Paths Not Taken: Understanding and Mending the Multilingual Factual Recall Pipeline

Meng Lu,Ruochen Zhang,Carsten Eickhoff,Ellie Pavlick
factual-recallmultilingualmechanistic-analysisDOIDBLP
6
泛读FindingsEMNLP 2025

Mitigating Hallucinations in Large Vision-Language Models by Self-Injecting Hallucinations

Yifan Lu,Ziqi Zhang,Chunfeng Yuan,Jun Gao,Congxuan Zhang,Xiaojuan Qi,Bing Li,Weiming Hu
vlmhallucinationpost-trainingAnthologyDBLP
6
泛读EMNLP 2025

Shared Path: Unraveling Memorization in Multilingual LLMs through Language Similarities

Xiaoyu Luo,Yiyi Chen,Johannes Bjerva,Qiongxiu Li
memorizationmultilingualmechanistic-analysisDOIDBLP
6
泛读FindingsEMNLP 2025

Towards Low-Resource Alignment to Diverse Perspectives with Sparse Feedback

Chu Fei Luo,Samuel Dahan,Xiaodan Zhu
alignmentrlhfsparse-feedbackAnthologyDBLP
6
泛读EMNLP 2025

Towards Robust Mathematical Reasoning

Thang Luong,Dawsen Hwang,Hoang H. Nguyen,Golnaz Ghiasi,Yuri Chervonyi,Insuk Seo ... 省略 10 位作者 ... ,Xingyou Song,Trieu H. Trinh,Quoc V. Le,Junehyuk Jung
math-reasoningrobustnesspost-trainingDOIDBLP
6
泛读EMNLP 2025

Recall with Reasoning: Chain-of-Thought Distillation for Mamba's Long-Context Memory and Extrapolation

Jun-Yu Ma,Tianqing Fang,Zhisong Zhang,Hongming Zhang,Haitao Mi,Dong Yu
mambalong-contextcot-distillationDOIDBLP
6
泛读EMNLP 2025

Cacheback: Speculative Decoding With Nothing But Cache

Zhiyao Ma,In Gim,Lin Zhong
speculative-decodingkv-cacheinferenceDOIDBLP
6
泛读EMNLP 2025

Large Language Models Have Intrinsic Meta-Cognition, but Need a Good Lens

LLM 经常产生幻觉,表面上看似乎缺乏对自己知识边界的认知(即元认知),但实际上这可能是因为缺乏有效的探测手段,导致内部置信度未被正确表达。

Ziyang Ma,Qingyue Yuan,Zhenglin Wang,Deyu Zhou
Southeast Universitymeta-cognitionevaluationreasoningDOIDBLP
6
泛读EMNLP 2025

You Are What You Train: Effects of Data Composition on Training Context-aware Machine Translation Models

训练上下文感知(Context-aware)的机器翻译模型时,句子级数据与文档级数据的配比对最终长文本翻译性能的影响机制尚不明确,容易导致模型变“懒”或基础能力下降。

Pawel Maka,Yusuf Can Semerci,Jan Scholtes,Gerasimos Spanakis
Maastricht Universitydata-mixturemachine-translationtraining-dynamicsDOIDBLP
6
泛读EMNLP 2025

All for One: LLMs Solve Mental Math at the Last Token With Information Transferred From Other Tokens

LLM 在没有外部草稿本(Scratchpad/CoT)的情况下进行心算,其内部 Transformer 注意力机制是如何完成信息聚合和隐式计算的,一直是个黑盒。

Siddarth Mamidanna,Daking Rai,Ziyu Yao,Yilun Zhou
MITUMass Amherstattentionreasoningtoken-interactionDOIDBLP
6
泛读EMNLP 2025

Pico: A Modular Framework for Hypothesis-Driven Small Language Model Research

这篇工作要解决的是:小语言模型研究缺少一个可复现实验、可快速改变量、并且能把“想验证的假设”明确落到训练配置上的统一框架。过去很多 SLM 结论来自一次性脚本、强耦合训练流水线或只适配单一模型族的代码,这使得数据、架构、优化器和 tokenizer 的作用很难被干净拆开比较。

Richard Diehl Martinez,David Demitri Africa,Yuval Weiss,Suchir Salhan,Ryan Daniels,Paula Buttery
small-language-modelresearch-frameworkpretrainingDOIDBLP
6
泛读EMNLP 2025

Demystifying optimized prompts in language models

这篇论文要解释的是:为什么优化后的 prompt 往往有效,以及它们到底在利用模型的什么机制。此前很多 prompt optimization 工作能找到高性能模板,但得到的是经验现象,难以判断这些 prompt 学到的是任务语义、格式偏置,还是某种脆弱的表面触发器。

Rimon Melamed,Lucas H. McCabe,H. Howie Huang
promptinterpretabilityoptimizationDOIDBLP
6
泛读IndustryEMNLP 2025

SFAL: Semantic-Functional Alignment Scores for Distributional Evaluation of Auto-Interpretability in Sparse Autoencoders

这篇工作要解决的是:如何更分布式、更可量化地评估稀疏自编码器(SAE)自动可解释性的质量,而不只看几个人工挑选 feature 的案例。过去 SAE 评估常卡在“看起来像解释了某些 feature”,但缺少同时覆盖语义一致性和功能一致性的系统指标。

Fabio Mercorio,Filippo Pallucchini,Daniele Potertì,Antonio Serino,Andrea Seveso
saeinterpretabilityevaluationDOIDBLP
6
泛读EMNLP 2025

Exploring the Hidden Capacity of LLMs for One-Step Text Generation

这篇论文探讨的是:LLM 是否具备被低估的一步生成能力,也就是不依赖逐 token 自回归展开,就能在单步中生成有用文本。这个问题值得重提,因为 AR 是当前主流范式,但它在推理延迟、并行性和全局一致性上有天然限制。

Gleb Mezentsev,Ivan V. Oseledets
generationcapacityanalysisDOIDBLP
6
泛读EMNLP 2025

Discursive Circuits: How Do Language Models Understand Discourse Relations?

这篇论文想回答的是:语言模型究竟如何表示和利用 discourse relations,也就是句间/段间的话语关系。以往模型能在篇章任务上得分,并不等于我们知道它是否真的学到了因果、转折、让步、展开等关系,还是只是在利用浅层连接词和局部模板。

Yisong Miao,Min-Yen Kan
circuitsdiscourseinterpretabilityDOIDBLP
6
泛读FindingsEMNLP 2025

Zipf's and Heaps' Laws for Tokens and LLM-generated Texts

这篇论文研究的是:Zipf 定律和 Heaps 定律在 token 序列与 LLM 生成文本中是否仍成立、如何偏离,以及这种偏离说明了什么。这个问题以前常停留在自然语言统计层面,但现在值得重做,因为 tokenizer 和生成模型改变了观测单位与生成机制,经典语言统计未必能直接沿用。

Nikolay Mikhaylovskiy
zipftokenizationanalysisAnthologyDBLP
6
泛读FindingsEMNLP 2025

Stratified Selective Sampling for Instruction Tuning with Dedicated Scoring Strategy

这篇论文解决的是:指令微调数据如何在预算有限时做更有效的筛选,而不是简单随机采样或只按单一质量分数取 top。过去 instruction tuning 的常见做法要么盲目堆数据,要么依赖一个粗糙打分器直接过滤,但这往往牺牲覆盖度,导致模型在长尾任务或困难样本上退化。

Paramita Mirza,Lucas Weber,Fabian Küch
instruction-tuningdata-selectionAnthologyDBLP
6
泛读FindingsEMNLP 2025

LimaCost: Data Valuation for Instruction Tuning of Large Language Models

Hyeonseok Moon,Jaehyung Seo,Seonmin Koo,Jinsung Kim,Young-kyoung Ham,Jiwon Moon,Heuiseok Lim
data-valuationinstruction-tuningAnthologyDBLP
6
泛读EMNLP 2025

Jailbreak-Tuning: Models Efficiently Learn Jailbreak Susceptibility

Brendan Murphy,Dillon Bowen,Shahrad Mohammadzadeh,Tom Tseng,Julius Broomfield,Adam Gleave,Kellin Pelrine
safetyjailbreakfine-tuningDOIDBLP
6
泛读EMNLP 2025

Controllable Memorization in LLMs via Weight Pruning

Chenjie Ni,Zhepeng Wang,Runxue Bao,Shangqian Gao,Yanfu Zhang
memorizationpruningDOIDBLP
6
泛读FindingsEMNLP 2025

Mechanistic Understanding and Mitigation of Language Confusion in English-Centric Large Language Models

Ercong Nie,Helmut Schmid,Hinrich Schütze
multilingualinterpretabilityAnthologyDBLP
6
泛读EMNLP 2025

Efficient Inference for Large Language Models -Algorithm, Model, and System

Xuefei Ning,Guohao Dai,Haoli Bai,Lu Hou,Yu Wang,Qun Liu
inferencesurveyDOIDBLP
6
泛读EMNLP 2025

SLlama: Parameter-Efficient Language Model Architecture for Enhanced Linguistic Competence Under Strict Data Constraints

Victor Adelakun Omolaoye,Babajide Alamu Owoyele,Gerard de Melo
parameter-efficientarchitecturedata-constraintDOIDBLP
6
泛读FindingsEMNLP 2025

A Survey on Training-free Alignment of Large Language Models

Birong Pan,Yongqi Li,Weiyu Zhang,Wenpeng Lu,Mayi Xu,Shen Zhou,Yuanyuan Zhu,Ming Zhong,Tieyun Qian
alignmenttraining-freesurveyAnthologyDBLP
6
泛读EMNLP 2025

PLAN-TUNING: Post-Training Language Models to Learn Step-by-Step Planning for Complex Problem Solving

Mihir Parmar,Palash Goyal,Xin Liu,Yiwen Song,Mingyang Ling,Chitta Baral,Hamid Palangi,Tomas Pfister
post-trainingplanningreasoningDOIDBLP
6
泛读EMNLP 2025

PlanGEN: A Multi-Agent Framework for Generating Planning and Reasoning Trajectories for Complex Problem Solving

Mihir Parmar,Xin Liu,Palash Goyal,Yanfei Chen,Long T. Le,Swaroop Mishra ... 省略 4 位作者 ... ,Chitta Baral,Chen-Yu Lee,Tomas Pfister,Hamid Palangi
multi-agentsynthetic-datareasoning-trajectoriesDOIDBLP
6
泛读IndustryEMNLP 2025

Harmonizing Diverse Models: A Layer-wise Merging Strategy for Consistent Generation

Xujun Peng,Anoop Kumar,Jingyu Wu,Parker Glenn,Daben Liu
model-merginglayer-wise-mergingpost-trainingDOIDBLP
6
泛读FindingsEMNLP 2025

Mixed Signals: Decoding VLMs' Reasoning and Underlying Bias in Vision-Language Conflict

Pouya Pezeshkpour,Moin Aminnaseri,Estevam Hruschka
vlmmultimodal-reasoningmodality-conflictAnthologyDBLP
6
泛读EMNLP 2025

Interpretability Analysis of Arithmetic In-Context Learning in Large Language Models

LLM 是如何通过上下文学习(ICL)完成算术推理的?其内部机制一直是个黑盒。

Gregory Polyakov,Christian Hepting,Carsten Eickhoff,Seyed Ali Bahrainian
interpretabilityiclarithmeticDOIDBLP
6
泛读FindingsEMNLP 2025

MidPO: Dual Preference Optimization for Safety and Helpfulness in Large Language Models via a Mixture of Experts Framework

在偏好优化(如 DPO)阶段,提升模型的安全性往往会以牺牲有用性(Helpfulness)为代价,产生严重的“对齐税”(Alignment Tax)。

Yupeng Qi,Ziyu Lyu,Min Yang,Yanlin Wang,Lu Bai,Lixin Cui
dposafetyhelpfulnessAnthologyDBLP
6
泛读EMNLP 2025

Data Drives Unstable Hierarchical Generalization in LMs

这篇工作要解决的是语言模型层级泛化不稳定这一现象到底来自模型归纳偏置,还是首先由数据驱动。过去很多关于 compositionality 或 hierarchical generalization 的讨论容易把失败归因于架构或训练目标,但数据分布、表面线索和采样不均衡可能已经足以制造看似深层的泛化差异。

Tian Qin,Naomi Saphra,David Alvarez-Melis
generalizationdata-mixturetraining-dynamicsDOIDBLP
6
泛读EMNLP 2025

Orthogonal Finetuning Made Scalable

这篇工作要解决的是 Orthogonal Finetuning 虽然参数高效、干扰小,但在大模型和实际训练设置下扩展性不足。过去这类正交约束方法常见的问题是优化开销高、实现不稳,或者当 rank 和层数上去后训练成本迅速膨胀,导致它们理论上干净、工程上难用。

Zeju Qiu,Weiyang Liu,Adrian Weller,Bernhard Schölkopf
finetuningparameter-efficientscalingDOIDBLP
6
泛读EMNLP 2025

ThinkTuning: Instilling Cognitive Reflections without Distillation

这篇工作要解决的是让模型学会认知反思式行为时,大家过度依赖强教师蒸馏。现有做法往往用更强模型生成 reflection、critique 或 revision 轨迹,再训练小模型模仿;问题是教师成本高、风格依赖重,而且学生容易只学到表面格式而不是何时该反思。

Aswin RRV,Jacob Dineen,Divij Handa,Md Nayem Uddin,Mihir Parmar,Chitta Baral,Ben Zhou
reasoningfinetuningdistillation-freeDOIDBLP
6
泛读FindingsEMNLP 2025

SeaPO: Strategic Error Amplification for Robust Preference Optimization of Large Language Models

这篇工作要解决的是 preference optimization 在噪声偏好、弱对比样本和分布外错误面前不够稳。标准 DPO/IPO 类方法通常把 chosen-rejected 对当成可靠监督,但一旦 rejected 太弱、错误模式过于单一,模型会学到脆弱边界,面对稍有变化的坏样本就容易失效。

Jun Rao,Yunjie Liao,Xuebo Liu,Zepeng Lin,Lian Lian,Dong Jin,Shengjun Cheng,Jun Yu,Min Zhang
dpopreference-optimizationAnthologyDBLP
6
泛读EMNLP 2025

Alignment with Fill-In-the-Middle for Enhancing Code Generation

Houxing Ren,Zimu Lu,Weikang Shi,Haotian Hou,Yunqiao Yang,Ke Wang,Aojun Zhou,Junting Pan,Mingjie Zhan,Hongsheng Li
codefimalignmentDOIDBLP
6
泛读FindingsEMNLP 2025

Sparsity May Be All You Need: Sparse Random Parameter Adaptation

Jesus Rios,Pierre L. Dognin,Ronny Luss,Karthikeyan Natesan Ramamurthy
sparsitypeftloraAnthologyDBLP
6
泛读FindingsEMNLP 2025

Krikri: Advancing Open Large Language Models for Greek

Dimitris Roussis,Leon Voukoutis,Georgios Paraskevopoulos,Sokratis Sofianopoulos,Prokopis Prokopidis,Vassilis P. Plagianakos,Athanasios Katsamanis,Stelios Piperidis,Vassilis Katsouros
multilingualgreekopen-modelAnthologyDBLP
6
泛读EMNLP 2025

BabyLM's First Constructions: Causal interventions provide a signal of learning

Joshua Rozner,Leonie Weissweiler,Cory Shain
babylmcausal-interventionlearning-dynamicsDOIDBLP
6
泛读EMNLP 2025

Instructing Large Language Models for Low-Resource Languages: A Systematic Study for Basque

Oscar Sainz,Naiara Pérez,Julen Etxaniz,Joseba Fernandez de Landa,Itziar Aldabe,Iker García-Ferrero ... 省略 2 位作者 ... ,German Rigau,Eneko Agirre,Mikel Artetxe,Aitor Soroa
low-resourceinstruction-tuningmultilingualDOIDBLP
6
泛读EMNLP 2025

Mixing Inference-time Experts for Enhancing LLM Reasoning

Soumya Sanyal,Tianyi Xiao,Xiang Ren
moeinference-timereasoningDOIDBLP
6
泛读EMNLP 2025

Mind the Gap: A Closer Look at Tokenization for Multiple-Choice Question Answering with LLMs

Mario Sanz-Guerrero,Minh Duc Bui,Katharina von der Wense
tokenizermcqaevaluationDOIDBLP
6
泛读FindingsEMNLP 2025

DIPLomA: Efficient Adaptation of Instructed LLMs to Low-Resource Languages via Post-Training Delta Merging

Ixak Sarasua,Ander Corral,Xabier Saralegi
low-resourcemodel-mergingdelta-mergingAnthologyDBLP
6
泛读FindingsEMNLP 2025

Characterizing Positional Bias in Large Language Models: A Multi-Model Evaluation of Prompt Order Effects

Patrick Schilcher,Dominik Karasin,Michael Schöpf,Haisam Saleh,Antonela Tommasel,Markus Schedl
positional-biasevaluationanalysisAnthologyDBLP
6
泛读EMNLP 2025

The Good, the Bad, and the Debatable: A Survey on the Impacts of Data for In-Context Learning

Stephanie Schoch,Yangfeng Ji
in-context-learningdatasurveyDOIDBLP
6
泛读FindingsEMNLP 2025

FLAMES: Improving LLM Math Reasoning via a Fine-Grained Analysis of the Data Synthesis Pipeline

Parker Seegmiller,Kartik Mehta,Soumya Saha,Chenyang Tao,Shereen Oraby,Arpit Gupta,Tagyoung Chung,Mohit Bansal,Nanyun Peng
data-synthesismath-reasoningsftAnthologyDBLP
6
泛读EMNLP 2025

Enhancing Speech Large Language Models with Prompt-Aware Mixture of Audio Encoders

Weiqiao Shan,Yuang Li,Yuhao Zhang,Yingfeng Luo,Chen Xu,Xiaofeng Zhao ... 省略 2 位作者 ... ,Min Zhang,Hao Yang,Tong Xiao,JingBo Zhu
speech-llmaudio-encodermixture-of-expertsDOIDBLP
6
泛读EMNLP 2025

CYCLE-INSTRUCT: Fully Seed-Free Instruction Tuning via Dual Self-Training and Cycle Consistency

现有指令微调依赖人工种子数据或强模型蒸馏来启动数据合成流程,这限制了可扩展性和自主性。本文要解决的是:能否在完全不需要种子指令数据的情况下,让模型自己生成高质量指令-回复对来完成 SFT?

Zhanming Shen,Hao Chen,Yulei Tang,Shaolin Zhu,Wentao Ye,Xiaomeng Hu,Haobo Wang,Gang Chen,Junbo Zhao
instruction-tuningself-trainingdata-synthesisDOIDBLP
6
泛读EMNLP 2025

CondenseLM: LLMs-driven Text Dataset Condensation via Reward Matching

大规模文本数据集的存储和训练成本高昂,数据集蒸馏/浓缩(dataset condensation)可以用小数据集替代大数据集训练,但现有方法主要面向图像分类,难以直接应用于文本和 LLM 场景。本文要解决的是如何用 LLM 驱动的方式对文本数据集进行浓缩。

Cheng Shen,Yew-Soon Ong,Joey Tianyi Zhou
dataset-condensationdata-qualityreward-matchingDOIDBLP
6
泛读EMNLP 2025

From Behavioral Performance to Internal Competence: Interpreting Vision-Language Models with VLM-Lens

VLM(视觉-语言模型)在行为层面(benchmark 分数)的表现不一定反映其内部真正的视觉理解能力。本文要解决的是如何从内部表示层面(而非仅看输出)来解释和评估 VLM 的视觉理解能力。

Hala Sheta,Eric Huang,Shuyu Wu,Ilia Alenabi,Jiajun Hong,Ryker Lin ... 省略 2 位作者 ... ,Jialin Yang,Jiawei Zhou,Ziqiao Ma,Freda Shi
vlminterpretabilitymechanism-analysisDOIDBLP
6
泛读FindingsEMNLP 2025

InsBank: Evolving Instruction Subset for Ongoing Alignment

持续对齐(ongoing alignment)场景下,指令数据不断增长,但并非所有数据都对当前模型状态有价值。如何动态选择最有效的指令子集来维持对齐效果,同时控制训练成本?

Jiayi Shi,Yiwei Li,Shaoxiong Feng,Peiwen Yuan,Xinglin Wang,Yueqi Zhang ... 省略 1 位作者 ... ,Boyuan Pan,Huan Ren,Yao Hu,Kan Li
instruction-tuningdata-selectionalignmentAnthologyDBLP
6
泛读FindingsEMNLP 2025

SpecCoT: Accelerating Chain-of-Thought Reasoning through Speculative Exploration

标准的自回归(AR)生成在处理长思维链(CoT)推理时推理延迟极高,而传统的推测解码(Speculative Decoding)由于难以用小模型精准预测复杂的逻辑轨迹,在 CoT 任务上接受率极低。

Junhan Shi,Yijia Zhu,Zhenning Shi,Dan Zhao,Qing Li,Yong Jiang
speculative-decodingcotinference-accelerationAnthologyDBLP
6
泛读EMNLP 2025

Efficient Unstructured Pruning of Mamba State-Space Models for Resource-Constrained Environments

Mamba 等状态空间模型(SSM)在端侧设备部署时受限于内存带宽,但现有的非结构化剪枝方法大多针对 Transformer 的注意力头设计,直接用于 SSM 会破坏其隐状态动力学。

Ibne Farabi Shihab,Sanjeda Akter,Anuj Sharma
mambamodel-pruningstate-space-modelsDOIDBLP
6
泛读FindingsEMNLP 2025

Trust Me, I'm Wrong: LLMs Hallucinate with Certainty Despite Knowing the Answer

LLM 在生成幻觉(Hallucination)时往往表现出极高的置信度,过去常认为这是因为模型预训练数据中缺乏相关知识,导致其“不知为不知”的能力缺失。

Adi Simhi,Itay Itzhak,Fazl Barez,Gabriel Stanovsky,Yonatan Belinkov
Technionhallucinationinternal-knowledgecalibrationAnthologyDBLP
6
泛读EMNLP 2025

Improving Large Language Model Safety with Contrastive Representation Learning

传统的安全对齐(如 SFT 或标准 RLHF)仅在输出端(Logits)惩罚有害回复,导致模型内部仍保留有害特征的表征路径,极易被越狱提示(Jailbreak)通过对抗扰动绕过。

Samuel Simko,Mrinmaya Sachan,Bernhard Schölkopf,Zhijing Jin
MPIETH Zurichsafety-alignmentrepresentation-learningcontrastive-learningDOIDBLP
6
泛读IndustryEMNLP 2025

Semantic Agreement Enables Efficient Open-Ended LLM Cascades

这篇工作解决的是开放式 LLM cascade 的效率问题:怎样在不显著损失答案质量的情况下,让便宜模型先做、贵模型后兜底。难点在于开放式生成没有像分类那样简单的置信度判据,传统 cascade 很容易把难题错误地下放给小模型,导致质量不稳定。

Duncan Soiffer,Steven Kolawole,Virginia Smith
cascadesevaluationefficiencyDOIDBLP
6
泛读FindingsEMNLP 2025

Saten: Sparse Augmented Tensor Networks for Post-Training Compression of Large Language Models

这篇工作要解决的是:如何在 post-training 阶段压缩大语言模型,同时尽量保住能力,而不必重新做高成本蒸馏或再训练。常见压缩手段如剪枝、低秩分解、量化各有代价:剪枝易伤精度,低秩对复杂权重不够灵活,量化偏重数值表示而非结构压缩;Saten 看起来是在这些路线之间找一个更强表达的压缩形式。

Ryan Solgi,Kai Zhen,Rupak Vignesh Swaminathan,Nathan Susanj,Athanasios Mouchtaris,Siegfried Kunzmann,Zheng Zhang
compressionpost-trainingsparsityAnthologyDBLP
6
泛读FindingsEMNLP 2025

Improving the Language Understanding Capabilities of Large Language Models Using Reinforcement Learning

这篇工作关注一个仍有争议的问题:强化学习能否实质性提升 LLM 的语言理解,而不仅是改善格式遵循、偏好对齐或推理搜索。过去 RL 在 LLM 上的主要成功更多体现在对齐和 reasoning style,真正的 language understanding 提升常被质疑只是 benchmark overfitting 或 response shaping。

Sai Ashish Somayajula,Bokai Hu,Qi Cao,Xin Pan,Pengtao Xie
reinforcement-learninglanguage-understandingalignmentAnthologyDBLP
6
泛读EMNLP 2025

Arena-lite: Efficient and Reliable Large Language Model Evaluation via Tournament-Based Direct Comparisons

这篇工作要解决的是大模型评测的效率和可靠性矛盾。当前 arena 式直接比较虽然贴近人类偏好,但配对数量随模型数增长很快,且评测方差受对手选择影响大;Arena-lite 试图用 tournament-based direct comparisons 在更少比较下保持稳定排序。

Seonil Son,Ju-Min Oh,Heegon Jin,Cheolhun Jang,Jeongbeom Jeong,Kuntae Kim
evaluationrankingarenaDOIDBLP
6
泛读FindingsEMNLP 2025

Both Text and Images Leaked! A Systematic Analysis of Data Contamination in Multimodal LLM

Dingjie Song,Sicheng Lai,Mingxuan Wang,Shunian Chen,Lichao Sun,Benyou Wang
data-contaminationmultimodalevaluationAnthologyDBLP
6
泛读EMNLP 2025

Unsupervised Hallucination Detection by Inspecting Reasoning Processes

Ponhvoan Srey,Xiaobao Wu,Anh Tuan Luu
hallucinationreasoningdetectionDOIDBLP
6
泛读EMNLP 2025

DEBATE, TRAIN, EVOLVE: Self-Evolution of Language Model Reasoning

Gaurav Srivastava,Zhenyu Bi,Meng Lu,Xuan Wang
self-playreasoningself-improvementDOIDBLP
6
泛读EMNLP 2025

ThinkSLM: Towards Reasoning in Small Language Models

Gaurav Srivastava,Shuxiang Cao,Xuan Wang
small-language-modelreasoningdistillationDOIDBLP
6
泛读FindingsEMNLP 2025

GmSLM : Generative Marmoset Spoken Language Modeling

Talia Sternberg,Michael London,David Omer,Yossi Adi
speech-lmanimal-languagegenerativeAnthologyDBLP
6
泛读FindingsEMNLP 2025

Toolscaler: Scalable Generative Tool Calling via Structure-Aware Semantic Tokenization

Yunyue Su,Zhang Jinshuai,Bowen Fang,Wen Ye,Jinghao Zhang,Bowen Song,Weiqiang Wang,Qiang Liu,Liang Wang
tool-usetokenizationgenerative-retrievalAnthologyDBLP
6
泛读EMNLP 2025

Training LLMs to be Better Text Embedders through Bidirectional Reconstruction

Chang Su,Dengliang Shi,Siyuan Huang,Jintao Du,Changhua Meng,Yu Cheng,Weiqiang Wang,Zhouhan Lin
embeddingbidirectionalcontrastiveDOIDBLP
6
泛读FindingsEMNLP 2025

Understanding How Value Neurons Shape the Generation of Specified Values in LLMs

Yi Su,Jiayi Zhang,Shu Yang,Xinhai Wang,Lijie Hu,Di Wang
neuronsvaluesinterpretabilityAnthologyDBLP
6
泛读EMNLP 2025

Personality Vector: Modulating Personality of Large Language Models by Model Merging

Seungjong Sun,Seo Yeon Baek,Jang Hyun Kim
model-mergingpersonalitysteeringDOIDBLP
6
泛读EMNLP 2025

Probing for Arithmetic Errors in Language Models

Yucheng Sun,Alessandro Stolfo,Mrinmaya Sachan
arithmeticprobinginterpretabilityDOIDBLP
6
泛读FindingsEMNLP 2025

SimpleDeepSearcher: Deep Information Seeking via Web-Powered Reasoning Trajectory Synthesis

Shuang Sun,Huatong Song,Yuhao Wang,Ruiyang Ren,Jinhao Jiang,Junjie Zhang ... 省略 3 位作者 ... ,Zheng Liu,Lei Fang,Zhongyuan Wang,Ji-Rong Wen
deep-searchreasoningdata-synthesisAnthologyDBLP
6
泛读EMNLP 2025

ThinkEdit: Interpretable Weight Editing to Mitigate Overly Short Thinking in Reasoning Models

Chung-En Sun,Ge Yan,Tsui-Wei Weng
weight-editingreasoninginterpretabilityDOIDBLP
6
泛读EMNLP 2025

Randomly Removing 50% of Dimensions in Text Embeddings has Minimal Impact on Retrieval and Classification Tasks

Sotaro Takeshita,Yurina Takeshita,Daniel Ruffinelli,Simone Paolo Ponzetto
embeddingdimensionalityretrievalDOIDBLP
6
泛读EMNLP 2025

Surprise Calibration for Better In-Context Learning

Zhihang Tan,Jingrui Hou,Ping Wang,Qibiao Hu,Peng Zhu
iclcalibrationDOIDBLP
6
泛读FindingsEMNLP 2025

Seeing is Believing: Emotion-Aware Audio-Visual Language Modeling for Expressive Speech Generation

Weiting Tan,Jiachen Lian,Hirofumi Inaguma,Paden Tomasello,Philipp Koehn,Xutai Ma
speech-lmaudio-visualmultimodalAnthologyDBLP
6
泛读EMNLP 2025

Too Consistent to Detect: A Study of Self-Consistent Errors in LLMs

Hexiang Tan,Fei Sun,Sha Liu,Du Su,Qi Cao,Xin Chen ... 省略 1 位作者 ... ,Xunliang Cai,Yuanzhuo Wang,Huawei Shen,Xueqi Cheng
hallucinationself-consistencyerror-analysisDOIDBLP
6
泛读FindingsEMNLP 2025

Decoupled Proxy Alignment: Mitigating Language Prior Conflict for Multimodal Alignment in MLLMs

Chenkun Tan,Pengyu Wang,Shaojun Zhou,Botian Jiang,Zhaowei Li,Dong Zhang,Xinghao Wang,Yaqian Zhou,Xipeng Qiu
mllmalignmentlanguage-priorAnthologyDBLP
6
泛读EMNLP 2025

Middo: Model-Informed Dynamic Data Optimization for Enhanced LLM Fine-Tuning via Closed-Loop Learning

Zinan Tang,Xin Gao,Qizhi Pei,Zhuoshi Pan,Mengzhang Cai,Jiang Wu,Conghui He,Lijun Wu
sftdata-qualitydata-selectionDOIDBLP
6
泛读FindingsEMNLP 2025

Large Language Models Might Not Care What You Are Saying: Prompt Format Beats Descriptions

Chenming Tang,Zhixiang Wang,Hao Sun,Yunfang Wu
prompt-formaticlanalysisAnthologyDBLP
6
泛读EMNLP 2025

Enhancing Chain-of-Thought Reasoning via Neuron Activation Differential Analysis

Yiru Tang,Kun Zhou,Yingqian Min,Wayne Xin Zhao,Jing Sha,Zhichao Sheng,Shijin Wang
cotneuron-analysisreasoningDOIDBLP
6
泛读EMNLP 2025

Language Models Can be Efficiently Steered via Minimal Embedding Layer Transformations

Diogo Tavares,David Semedo,Alexander Rudnicky,João Magalhães
steeringembeddingrepresentationDOIDBLP
6
泛读EMNLP 2025

Prior Prompt Engineering for Reinforcement Fine-Tuning

Pittawat Taveekitworachai,Potsawee Manakul,Sarana Nutanong,Kunat Pipatanakul
rlftprompt-engineeringDOIDBLP
6
泛读FindingsEMNLP 2025

MoVoC: Morphology-Aware Subword Construction for Ge'ez Script Languages

Hailay Kidu Teklehaymanot,Dren Fazlija,Wolfgang Nejdl
tokenizersubwordlow-resourceAnthologyDBLP
6
泛读EMNLP 2025

FaST: Feature-aware Sampling and Tuning for Personalized Preference Alignment with Limited Data

在极少量的用户特定数据下,如何对 LLM 进行个性化偏好对齐而不引发过拟合或灾难性遗忘。

Thibaut Thonet,Germán Kruszewski,Jos Rozen,Pierre Erbacher,Marc Dymetman
Naver Labs Europealignmentpreferencedata-efficientDOIDBLP
6
泛读FindingsEMNLP 2025

Towards Reverse Engineering of Language Models: A Survey

闭源 LLM 的黑盒特性导致难以评估其训练数据、架构和安全性,缺乏系统性的逆向工程分析框架。

Xinpeng Ti,Wentao Ye,Zhifang Zhang,Junbo Zhao,Chang Yao,Lei Feng,Haobo Wang
Zhejiang Universitysurveyreverse-engineeringinterpretabilityAnthologyDBLP
6
泛读FindingsEMNLP 2025

Context Copying Modulation: The Role of Entropy Neurons in Managing Parametric and Contextual Knowledge Conflicts

当上下文信息(如 RAG 注入的知识)与模型预训练的参数化知识发生冲突时,LLM 内部是如何进行路由和决策的。

Zineddine Tighidet,Andrea Mogini,Hédi Ben-Younes,Jiali Mei,Patrick Gallinari,Benjamin Piwowarski
Sorbonne UniversityCriteointerpretabilityknowledge-conflictentropy-neuronsAnthologyDBLP
6
泛读FindingsEMNLP 2025

Internal states before wait modulate reasoning patterns

引入“等待/暂停 Token”(Wait/Pause Tokens)以增加推理计算量时,模型内部的推理轨迹是如何被改变的。

Dmitrii Troitskii,Koyena Pal,Chris Wendler,Callum McDougall
Apollo Researchreasoninginternal-stateswait-tokensAnthologyDBLP
6
泛读OutstandingEMNLP 2025

Measuring Chain of Thought Faithfulness by Unlearning Reasoning Steps

LLM 生成的思维链(CoT)究竟是驱动最终答案的真实因果推理过程(Faithfulness),还是仅仅是事后合理化(Post-hoc rationalization)。

Martin Tutek,Fateme Hashemi Chaleshtori,Ana Marasovic,Yonatan Belinkov
Technionchain-of-thoughtfaithfulnessunlearningDOIDBLP
6
泛读EMNLP 2025

CodeSSM: Towards State Space Models for Code Understanding

这篇论文要解决的是:状态空间模型(SSM)在长序列上很高效,但它们是否适合代码理解还缺少系统验证。代码和自然语言不同,依赖结构更离散、更层级化,很多人默认 attention 更适合处理这种非局部关系,因此 SSM 在代码上的位置一直不清楚。

Shweta Verma,Abhinav Anand,Mira Mezini
state-space-modelcode-understandingssmDOIDBLP
6
泛读EMNLP 2025

Let's Reason Formally: Natural-Formal Hybrid Reasoning Enhances LLM's Math Capability

这篇论文要解决的是:纯自然语言链式推理在数学任务上容易漂移,纯 formal 方法又难以覆盖开放式问题表达,因此需要一种更稳的自然语言—形式语言混合推理路径。过去很多工作在两端摇摆:要么全靠 CoT,要么让模型直接生成 Lean/程序;两者都各有脆弱点。

Ruida Wang,Yuxin Li,Yi R. Fung,Tong Zhang
reasoningmathformal-reasoningDOIDBLP
6
泛读EMNLP 2025

Model Unlearning via Sparse Autoencoder Subspace Guided Projections

如何在不损伤模型整体能力的前提下做精准 unlearning。现有方法(梯度上升、任务算术等)要么忘不干净,要么会把相邻能力一起抹掉,根源是它们在参数空间做 coarse 编辑,没有抓住'待忘知识'到底占据了哪个子空间。

Xu Wang,Zihao Li,Benyou Wang,Yan Hu,Difan Zou
香港中文大学(深圳)香港大学unlearningsparse-autoencoderrepresentationDOIDBLP
6
泛读EMNLP 2025

QCRD: Quality-guided Contrastive Rationale Distillation for Large Language Models

CoT 蒸馏里教师生成的 rationale 质量参差不齐,把好的坏的一股脑塞给学生会引入噪声。作者要的是能按质量区别对待的蒸馏方式。

Wei Wang,Zhaowei Li,Qi Xu,Yiqing Cai,Hang Song,Qi Qi,Ran Zhou,Zhida Huang,Tao Wang,Li Xiao
distillationrationalereasoningDOIDBLP
6
泛读EMNLP 2025

Token-Aware Editing of Internal Activations for Large Language Model Alignment

做 LLM 对齐时,全局地改激活(比如 activation steering、ITI 类方法)容易伤通用能力,因为不同 token 对'是否需要纠偏'的需求完全不同。

Tianbo Wang,Yuqing Ma,Kewei Liao,Chengzhao Yang,Zhange Zhang,Jiakai Wang,Xianglong Liu
activation-editingalignmentrepresentationDOIDBLP
6
泛读EMNLP 2025

Unleashing the Reasoning Potential of LLMs by Critique Fine-Tuning on One Problem

能不能用极少的数据把 base model 的推理能力'激活'出来——作者直接推到极端:就一道题,用 critique fine-tuning 能走多远。这是在 R1-Zero 范式之后对'推理能力是激发不是教会'这条思路的进一步压力测试。

Yubo Wang,Ping Nie,Kai Zou,Lijun Wu,Wenhu Chen
University of Waterlooreasoningcritiquefine-tuningDOIDBLP
6
泛读EMNLP 2025

Knowledge Editing through Chain-of-Thought

Changyue Wang,Weihang Su,Qingyao Ai,Yichen Tang,Yiqun Liu
knowledge-editingcotmodel-editingDOIDBLP
6
泛读FindingsEMNLP 2025

Token Knowledge: A New Perspective For Knowledge in Large Language Models

Jieyong Wang,Chunyao Song,Tingjian Ge
tokenknowledgerepresentationAnthologyDBLP
6
泛读EMNLP 2025

RewardDS: Privacy-Preserving Fine-Tuning for Large Language Models via Reward Driven Data Synthesis

Jianwei Wang,Chengming Shi,Junyao Yang,Haoran Li,Qianli Ma,Huiping Zhuang,Cen Chen,Ziqian Zeng
privacydata-synthesisreward-modelDOIDBLP
6
泛读EMNLP 2025

Alignment-Augmented Speculative Decoding with Alignment Sampling and Conditional Verification

Jikai Wang,Zhenxu Tian,Juntao Li,Qingrong Xia,Xinyu Duan,Zhe-Feng Wang,Baoxing Huai,Min Zhang
speculative-decodinginferenceDOIDBLP
6
泛读FindingsEMNLP 2025

Demystifying Multilingual Reasoning in Process Reward Modeling

Weixuan Wang,Minghao Wu,Barry Haddow,Alexandra Birch
prmmultilingualreasoningAnthologyDBLP
6
泛读EMNLP 2025

POSITION BIAS MITIGATES POSITION BIAS: Mitigate Position Bias Through Inter-Position Knowledge Distillation

Yifei Wang,Feng Xiong,Yong Wang,Linjing Li,Xiangxiang Chu,Daniel Dajun Zeng
position-biasdistillationDOIDBLP
6
泛读EMNLP 2025

Direct Judgement Preference Optimization

Peifeng Wang,Austin Xu,Yilun Zhou,Caiming Xiong,Shafiq Joty
dpojudgepreferenceDOIDBLP
6
泛读EMNLP 2025

Examining False Positives under Inference Scaling for Mathematical Reasoning

Yu Wang,Nan Yang,Liang Wang,Furu Wei,Fuli Feng
reasoninginference-scalingmathDOIDBLP
6
泛读FindingsEMNLP 2025

Faster and Better LLMs via Latency-Aware Test-Time Scaling

Zili Wang,Tianyu Zhang,Haoli Bai,Lu Hou,Xianzhi Yu,Wulong Liu,Shiming Xiang,Lei Zhu
test-time-scalinglatencyAnthologyDBLP
6
泛读FindingsEMNLP 2025

Under the Shadow of Babel: How Language Shapes Reasoning in LLMs

Chenxi Wang,Yixuan Zhang,Lang Gao,Zixiang Xu,Zirui Song,Yanbo Wang,Xiuying Chen
multilingualreasoningAnthologyDBLP
6
泛读EMNLP 2025

Weight-Aware Activation Sparsity with Constrained Bayesian Optimization Scheduling for Large Language Models

Ming Wang,Miao Zhang,Xuebo Liu,Liqiang Nie
activation-sparsitybayesianDOIDBLP
6
泛读FindingsEMNLP 2025

Visual Self-Refinement for Autoregressive Models

Jiamian Wang,Ziqi Zhou,Chaithanya Kumar Mummadi,Sohail A. Dianat,Majid Rabbani,Raghuveer Rao,Chen Qiu,Zhiqiang Tao
autoregressivevisual-generationself-refineAnthologyDBLP
6
泛读FindingsEMNLP 2025

ClusterUCB: Efficient Gradient-Based Data Selection for Targeted Fine-Tuning of LLMs

针对特定任务微调(Targeted SFT)时,基于梯度的核心集选择(Coreset Selection)计算开销过大,难以在 LLM 的数据规模下应用。

Zige Wang,Qi Zhu,Fei Mi,Minghui Xu,Ruochun Jin,Wenjing Yang
Tencent AI Labdata-selectionfine-tuninggradientAnthologyDBLP
6
泛读EMNLP 2025

UnifiedVisual: A Framework for Constructing Unified Vision-Language Datasets

视觉-语言预训练数据来源繁杂、格式不一,缺乏一个统一的框架来标准化构建高质量的多模态数据集。

Pengyu Wang,Shaojun Zhou,Chenkun Tan,Xinghao Wang,Wei Huang,Zhen Ye,Zhaowei Li,Botian Jiang,Dong Zhang,Xipeng Qiu
vlmdatasetvision-languageDOIDBLP
6
泛读FindingsEMNLP 2025

Attribution and Application of Multiple Neurons in Multimodal Large Language Models

多模态大模型(MLLMs)内部神经元对特定多模态概念或行为的协同归因机制尚不清晰,单神经元分析无法解释复杂的跨模态交互。

Feiyu Wang,Ziran Zhao,Dong Yu,Pengyuan Liu
Tencent AI LabinterpretabilityneuronsvlmAnthologyDBLP
6
泛读EMNLP 2025

JI2S: Joint Influence-Aware Instruction Data Selection for Efficient Fine-Tuning

指令微调数据选择中,单独评估每个样本的影响力(Influence)会忽略样本间的相互作用和冗余性,导致选出的数据同质化严重。

Jingyu Wei,Bo Liu,Tianjiao Wan,Baoyun Peng,Xingkong Ma,Mengmeng Guo
data-selectioninstruction-tuninginfluenceDOIDBLP
6
泛读EMNLP 2025

WebAgent-R1: Training Web Agents via End-to-End Multi-Turn Reinforcement Learning

现有的 Web Agent 多依赖监督微调(SFT),难以在真实、动态的网页环境中完成复杂的多步交互。

Zhepei Wei,Wenlin Yao,Yao Liu,Weizhi Zhang,Qin Lu,Liang Qiu ... 省略 2 位作者 ... ,Chao Zhang,Bing Yin,Hyokun Yun,Lihong Li
Tencent AI Labrlweb-agentmulti-turnDOIDBLP
6
泛读EMNLP 2025

Stop Looking for "Important Tokens" in Multimodal Language Models: Duplication Matters More

现有的视觉 token 压缩方法致力于寻找并保留“重要”的 token,但这种基于重要性打分的策略计算复杂且容易破坏空间上下文。

Zichen Wen,Yifeng Gao,Shaobo Wang,Junyuan Zhang,Qintong Zhang,Weijia Li,Conghui He,Linfeng Zhang
vlmtoken-pruningefficientDOIDBLP
6
泛读EMNLP 2025

reWordBench: Benchmarking and Improving the Robustness of Reward Models with Transformed Inputs

这篇论文要解决的是:reward model 对输入表面变换是否稳健,尤其是当语义基本不变但措辞、格式或结构被改写时,评分会不会漂移。过去很多 RM 评测默认输入分布干净且风格稳定,因此高分模型可能只是学到浅层偏好,而不是稳定的人类价值判断。

Zhaofeng Wu,Michihiro Yasunaga,Andrew Cohen,Yoon Kim,Asli Celikyilmaz,Marjan Ghazvininejad
reward-modelrobustnessbenchmarkDOIDBLP
6
泛读FindingsEMNLP 2025

ROSE: A Reward-Oriented Data Selection Framework for LLM Task-Specific Instruction Tuning

这篇论文要解决的是:任务特定 instruction tuning 时,训练数据很多但质量不均,如何选出最有利于最终任务奖励的数据。传统做法常用 embedding 相似度、困惑度过滤或人工规则,这些指标和最终 reward 并不总一致,因此容易选到看起来干净但对任务收益有限的数据。

Yang Wu,Huayi Zhang,Yizheng Jiao,Lin Ma,Xiaozhong Liu,Jinhong Yu,Dongyu Zhang,Dezhi Yu,Wei Xu
data-selectioninstruction-tuningreward-modelAnthologyDBLP
6
泛读EMNLP 2025

TokenSkip: Controllable Chain-of-Thought Compression in LLMs

这篇论文要解决的是:Chain-of-Thought 很有用,但推理 token 太贵,如何在尽量不伤答案质量的前提下压缩 CoT。过去常见做法是直接蒸馏短答案或用摘要式压缩思路,但这会失去可控性,用户无法指定压缩程度,也难在成本与正确率之间做精细权衡。

Heming Xia,Chak Tou Leong,Wenjie Wang,Yongqi Li,Wenjie Li
chain-of-thoughtcompressioninference-efficiencyDOIDBLP
6
泛读FindingsEMNLP 2025

DAPE-BR: Distance-Aware Positional Encoding for Mitigating Object Hallucination in LVLMs

LVLM 的 object hallucination 有一部分来自位置编码——图像 token 被摊平成一维序列后,视觉上相邻的 patch 在 PE 下可能很远,模型对视觉空间关系建模有偏,导致幻觉。

Mingrui Xie,Tianxiang Xu,Qianhai Tang,Shanming Yao,Xiaofeng Zhang,Junliang Du
vlmpositional-encodinghallucinationAnthologyDBLP
6
泛读FindingsEMNLP 2025

Beyond Surface Alignment: Rebuilding LLMs Safety Mechanism via Probabilistically Ablating Refusal Direction

现有安全对齐主要靠 SFT/RLHF 在行为层面抑制有害输出,但已有工作发现 refusal 行为可以被单个方向向量抹掉(refusal direction ablation 越狱)。作者想把安全机制做得更 robust,而不是只停留在表面。

Yuanbo Xie,Yingjie Zhang,Tianyun Liu,Duohe Ma,Tingwen Liu
safetyrepresentation-engineeringalignmentAnthologyDBLP
6
泛读EMNLP 2025

Re-Align: Aligning Vision Language Models via Retrieval-Augmented Direct Preference Optimization

VLM 做 DPO 对齐时,偏好数据(chosen vs rejected)一般靠 LLM 或人工构造,容易引入和视觉内容无关的偏好噪声,导致'语言偏见'越对越强。作者想让 DPO 更对齐'视觉事实'。

Shuo Xing,Peiran Li,Yuping Wang,Ruizheng Bai,Yueqi Wang,Chan-Wei Hu,Chengxuan Qian,Huaxiu Yao,Zhengzhong Tu
vlmdpoalignmentDOIDBLP
6
泛读EMNLP 2025

Mapping the Minds of LLMs: A Graph-Based Analysis of Reasoning LLMs

推理模型的 long CoT 到底长什么结构?是线性推理、树状探索还是有回溯?现有分析多停留在文本层面或 token 级统计,缺少结构化视角。

Zhen Xiong,Yujun Cai,Zhecheng Li,Yiwei Wang
reasoningmechanism-analysisgraph-basedDOIDBLP
6
泛读FindingsEMNLP 2025

The Progress Illusion: Revisiting meta-evaluation standards of LLM evaluators

Tianruo Rose Xu,Vedant Gaur,Liu Leqi,Tanya Goyal
evaluationllm-as-a-judgemeta-evaluationAnthologyDBLP
6
泛读Best PaperEMNLP 2025

Infini-gram mini: Exact n-gram Search at the Internet Scale with FM-Index

Hao Xu,Jiacheng Liu,Yejin Choi,Noah A. Smith,Hannaneh Hajishirzi
n-gramsearchdata-processingDOIDBLP
6
泛读FindingsEMNLP 2025

Structure Trumps Size: Rethinking Data Quality for LLM Reasoning

Hu Xu,Zeyan Li,Rui Wang,Jianfeng Xu
data-qualityreasoningsynthetic-dataAnthologyDBLP
6
泛读EMNLP 2025

DCR: Quantifying Data Contamination in LLMs Evaluation

Cheng Xu,Nan Yan,Shuhao Guan,Changhong Jin,Yuke Mei,Yibing Guo,M. Tahar Kechadi
data-contaminationevaluationbenchmarkingDOIDBLP
6
泛读FindingsEMNLP 2025

TwT: Thinking without Tokens by Habitual Reasoning Distillation with Multi-Teachers' Guidance

Jingxian Xu,Mengyu Zhou,Weichang Liu,Hanbing Liu,Shi Han,Dongmei Zhang
reasoning-distillationlatent-reasoningcotAnthologyDBLP
6
泛读FindingsEMNLP 2025

Addition in Four Movements: Mapping Layer-wise Information Trajectories in LLMs

Yao Yan
interpretabilitylayer-analysisarithmeticAnthologyDBLP
6
泛读EMNLP 2025

Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries

Tianyi Lorena Yan,Robin Jia
factualitygeneration-dynamicsknowledge-retrievalDOIDBLP
6
泛读FindingsEMNLP 2025

Towards Efficient CoT Distillation: Self-Guided Rationale Selector for Better Performance with Fewer Rationales

Jianzhi Yan,Le Liu,Youcheng Pan,Shiwei Chen,Yang Xiang,Buzhou Tang
cotdistillationrationaleAnthologyDBLP
6
泛读EMNLP 2025

Do Large Language Models Truly Grasp Addition? A Rule-Focused Diagnostic Using Two-Integer Arithmetic

Yang Yan,Yu Lu,Renjun Xu,Zhenzhong Lan
arithmeticgeneralizationdiagnosticDOIDBLP
6
泛读EMNLP 2025

Towards Holistic Evaluation of Large Audio-Language Models: A Comprehensive Survey

音频-语言大模型(ALMs)发展迅速,但缺乏一个统一、全面的评估框架,导致不同模型在语音识别、音频理解和跨模态推理上的能力难以横向对比。

Chih-Kai Yang,Neo S. Ho,Hung-yi Lee
National Taiwan University (NTU)audio-lmsurveyevaluationDOIDBLP
6
泛读FindingsEMNLP 2025

How Well Can Reasoning Models Identify and Recover from Unhelpful Thoughts?

随着自回归推理模型(如 o1 范式)的兴起,模型能否在生成长思维链时,自主识别出“无用的想法(死胡同)”并进行回溯纠错,成为了一个关键的机制盲区。

Sohee Yang,Sang-Woo Lee,Nora Kassner,Daniela Gottesman,Sebastian Riedel,Mor Geva
UCLCoherereasoningself-correctionAnthologyDBLP
6
泛读EMNLP 2025

MAIN: Mutual Alignment Is Necessary for instruction tuning

传统的指令微调(SFT)通常是单向的:强迫模型去拟合人类提供的目标回答。这会导致“知识冲突”——当目标回答超出模型预训练知识边界时,强行拟合会引发幻觉或灾难性遗忘。

Fanyi Yang,Jianfeng Liu,Xin Zhang,Haoyu Liu,Xixin Cao,Yuefeng Zhan,Hao Sun,Weiwei Deng,Feng Sun,Qi Zhang
instruction-tuningalignmentDOIDBLP
6
泛读EMNLP 2025

Process-Supervised Reinforcement Learning for Code Generation

Yufan Ye,Ting Zhang,Wenbin Jiang,Hua Huang
rlcodeprocess-rewardDOIDBLP
6
泛读EMNLP 2025

Towards Faithful Natural Language Explanations: A Study Using Activation Patching in Large Language Models

Wei Jie Yeo,Ranjan Satapathy,Erik Cambria
activation-patchingfaithfulnessinterpretabilityDOIDBLP
6
泛读EMNLP 2025

SEA: Supervised Embedding Alignment for Token-Level Visual-Textual Integration in MLLMs

Yuanyang Yin,Yaqi Zhao,Yajie Zhang,Yuanxing Zhang,Ke Lin,Jiahao Wang,Xin Tao,Pengfei Wan,Wentao Zhang,Feng Zhao
mllmvisual-tokensalignmentDOIDBLP
6
泛读EMNLP 2025

Back Attention: Understanding and Enhancing Multi-Hop Reasoning in Large Language Models

Zeping Yu,Yonatan Belinkov,Sophia Ananiadou
multi-hop-reasoningattention-analysismechanistic-interpretabilityDOIDBLP
6
泛读FindingsEMNLP 2025

Long-context Language Models Fail in Basic Retrieval Tasks Without Sufficient Reasoning Steps

Yijiong Yu,Zhixiao Qi,Yongfeng Huang,Wei Wang,Weifeng Liu,Ran Chen,Ji Pei
long-contextretrievalreasoningAnthologyDBLP
6
泛读IndustryEMNLP 2025

Z1: Efficient Test-time Scaling with Code

Zhaojian Yu,Yinghao Wu,Yilun Zhao,Arman Cohan,Xiao-Ping Zhang
test-time-scalingcode-reasoninginference-efficiencyDOIDBLP
6
泛读EMNLP 2025

Sequential-NIAH: A Needle-In-A-Haystack Benchmark for Extracting Sequential Needles from Long Contexts

Yifei Yu,Qian-Wen Zhang,Lingfeng Qiao,Di Yin,Fang Li,Jie Wang,Chen Zeng Xi,Suncong Zheng,Xiaolong Liang,Xing Sun
long-contextbenchmarkneedle-in-haystackDOIDBLP
6
泛读EMNLP 2025

Superficial Self-Improved Reasoners Benefit from Model Merging

Xiangchi Yuan,Chunhui Zhang,Zheyuan Liu,Dachuan Shi,Leyan Pan,Soroush Vosoughi,Wenke Lee
model-mergingself-improvementreasoningDOIDBLP
6
泛读FindingsEMNLP 2025

The Price of Format: Diversity Collapse in LLMs

Longfei Yun,Chenyang An,Zilong Wang,Letian Peng,Jingbo Shang
diversity-collapseformat-biasgeneration-diversityAnthologyDBLP
6
泛读IndustryEMNLP 2025

Beyond Dynamic Quantization: An Efficient Static Hierarchical Mix-precision Framework for Near-Lossless LLM Compression

Yi Zhang,Kai Zhang,Zheyang Li,Wenming Tan,Ye Ren,Jilin Hu
quantizationmodel-compressionmix-precisionDOIDBLP
6
泛读EMNLP 2025

Pre-training CLIP against Data Poisoning with Optimal Transport-based Matching and Alignment

CLIP 等视觉-语言预训练模型高度依赖海量网络图文对,极易受到数据投毒(如恶意错配图文)的攻击,导致 InfoNCE 损失传播错误梯度。

Tong Zhang,Kuofeng Gao,Jiawang Bai,Leo Yu Zhang,Xin Yin,Zonghui Wang,Shouling Ji,Wenzhi Chen
clipvlm-pretrainingdata-poisoningDOIDBLP
6
泛读FindingsEMNLP 2025

Reward Mixology: Crafting Hybrid Signals for Reinforcement Learning Driven In-Context Learning

使用强化学习优化 LLM 的上下文学习(ICL)能力时,单一的任务奖励信号过于稀疏且容易导致模型产生格式崩溃或奖励黑客(Reward Hacking)行为。

Changshuo Zhang,Ang Gao,Xiao Zhang,Yong Liu,Deyang Li,Fangchao Liu,Xinyu Zhang
rlin-context-learningreward-designAnthologyDBLP
6
泛读EMNLP 2025

LeanK: Learnable K Cache Channel Pruning for Efficient Decoding

长文本解码时 KV Cache 占用大量显存,现有压缩方法多集中于 Token 维度的丢弃(序列剪枝),忽视了 Key Cache 在通道(特征维度)上的冗余。

Yike Zhang,Zhiyuan He,Huiqiang Jiang,Chengruidong Zhang,Yuqing Yang,Jianyong Wang,Lili Qiu
MSRAkv-cachepruningefficient-decodingDOIDBLP
6
泛读EMNLP 2025

AdaptThink: Reasoning Models Can Learn When to Think

当前的推理模型(如 o1 或强制 CoT 的模型)对所有查询都应用长思维链,导致在简单任务上浪费大量算力并增加推理延迟。

Jiajie Zhang,Nianyi Lin,Lei Hou,Ling Feng,Juanzi Li
Tsinghua Universityreasoningtest-time-computeadaptive-computeDOIDBLP
6
泛读FindingsEMNLP 2025

CM-Align: Consistency-based Multilingual Alignment for Large Language Models

大模型的对齐(RLHF/DPO)通常在英文数据上进行,直接翻译偏好数据会导致质量下降,而多语言对齐往往面临“对齐税”,即非英语语言的性能或安全性受损。

Xue Zhang,Yunlong Liang,Fandong Meng,Songming Zhang,Yufeng Chen,Jinan Xu,Jie Zhou
TencentmultilingualalignmentconsistencyAnthologyDBLP
6
泛读FindingsEMNLP 2025

Test-Time Steering for Lossless Text Compression via Weighted Product of Experts

这篇工作解决的是:如何在不改模型参数的前提下,用测试时组合多个专家分布来提升无损文本压缩。传统做法通常依赖单一语言模型或训练期集成,压缩收益受限于模型偏好固定,而测试时如果能按上下文动态混合专家,可能拿到更低的编码代价。

Qihang Zhang,Muchen Li,Ziao Wang,Renjie Liao,Lele Wang
test-timeproduct-of-expertscompressionAnthologyDBLP
6
泛读EMNLP 2025

Exploring Chain-of-Thought Reasoning for Steerable Pluralistic Alignment

这篇工作讨论的是:在 pluralistic alignment 里,CoT 是否能帮助模型在多种合法价值取向之间做可控转向,而不是只学一个平均立场。现有对齐通常追求单一偏好最优解,容易把多元价值压平;而 steerable alignment 需要模型既能遵循特定立场,又不丢失推理过程,这正是 CoT 可能发挥作用的地方。

Yunfan Zhang,Kathleen McKeown,Smaranda Muresan
Columbia Universitychain-of-thoughtalignmentpreference-optimizationDOIDBLP
6
泛读EMNLP 2025

Shallow Focus, Deep Fixes: Enhancing Shallow Layers Vision Attention Sinks to Alleviate Hallucination in LVLMs

这篇工作要解决的是 LVLM hallucination 的一个具体来源:视觉注意力在浅层就出现 attention sinks,导致模型过早把注意力锁到少数无效位置,后续层再强也很难纠正。已有缓解幻觉的方法多在解码、对齐或后验校正层面动手,较少直接处理视觉流早期的信息路由失真。

Xiaofeng Zhang,Yihao Quan,Chen Shen,Chaochen Gu,Xiaosong Yuan,Shaotian Yan,Jiawei Cao,Hao Cheng,Kaijie Wu,Jieping Ye
lvlmhallucinationattentionDOIDBLP
6
泛读EMNLP 2025

CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling

这篇工作要解决的是:如何把 CLIP 有效地做成 MoE,而不是简单复制 ViT 或 LLM 里的稀疏专家套路。CLIP 的难点在于它是对比学习双塔结构,视觉和文本侧的对齐依赖共享嵌入空间,直接上 MoE 很容易破坏跨模态一致性或让专家退化成同质复制。

Jihai Zhang,Xiaoye Qu,Tong Zhu,Yu Cheng
clipmoeupcyclingDOIDBLP
6
泛读IndustryEMNLP 2025

A Reasoner for Real-World Event Detection: Scaling Reinforcement Learning via Adaptive Perplexity-Aware Sampling Strategy

这篇工作解决的是:在真实事件检测这类高噪声、长尾且反馈昂贵的任务上,如何把 RL 训练有效 scale 起来。常规 RL 数据采样往往对样本难度和模型不确定性不敏感,容易把算力浪费在过易或过难样本上,导致 reasoner 学不到稳定策略。

Xiaoyun Zhang,Jingqing Ruan,Xing Ma,Yawen Zhu,Jiansong Chen,Ke Zeng,Xunliang Cai
reinforcement-learningsamplingreasoningDOIDBLP
6
泛读FindingsEMNLP 2025

When to Continue Thinking: Adaptive Thinking Mode Switching for Efficient Reasoning

这篇工作关注推理效率里的一个核心决策:模型什么时候该继续思考,什么时候该直接输出答案。现有长推理方法往往统一给固定 thinking budget,简单但浪费,因为不同样本难度差异很大;而完全自由生成又容易出现过度思考和时延失控。

Xiaoyun Zhang,Jingqing Ruan,Xing Ma,Yawen Zhu,Haodong Zhao,Hao Li,Jiansong Chen,Ke Zeng,Xunliang Cai
adaptive-computereasoningtest-time-computeAnthologyDBLP
6
泛读EMNLP 2025

Membership and Memorization in LLM Knowledge Distillation

这篇工作要回答的是:LLM 知识蒸馏会如何影响 membership inference 和 memorization 风险。过去大家常把蒸馏看成压缩与泛化手段,默认学生模型会更少记忆训练样本,但这一点在大模型知识转移里并没有被充分验证,尤其当教师本身已携带可恢复的训练痕迹时。

Ziqi Zhang,Ali Shahin Shamsabadi,Hanxiao Lu,Yifeng Cai,Hamed Haddadi
knowledge-distillationmemorizationdata-leakageDOIDBLP
6
泛读FindingsEMNLP 2025

ISACL: Internal State Analyzer for Copyrighted Training Data Leakage

这篇工作要解决的是:如何检测模型是否泄露了受版权保护的训练数据,而且不只看输出复现,还看内部状态是否携带可识别痕迹。传统方法多依赖 prompt 诱导或字符串匹配,能抓到显式复述,但对隐式记忆、近似重构和未触发的内部存储不敏感。

Guangwei Zhang,Qisheng Su,Jiateng Liu,Cheng Qian,Yanzhou Pan,Yanjie Fu,Denghui Zhang
data-leakagememorizationtraining-dataAnthologyDBLP
6
泛读FindingsEMNLP 2025

Can We Steer Reasoning Direction by Thinking Intervention?

现在影响 LLM 推理方向基本靠改 prompt 或加 CoT 示例,作者问:能不能在模型内部(激活层面)直接干预 thinking 过程的方向?类似 representation engineering,但针对 reasoning。

Xingsheng Zhang,Luxi Xing,Chen Zhang,Yanbing Liu,Yifan Deng,Yunpeng Li,Yue Hu,Chenxu Niu
reasoningsteeringinterventionAnthologyDBLP
6
泛读EMNLP 2025

LeTS: Learning to Think-and-Search via Process-and-Outcome Reward Hybridization

Agentic search 场景里训练 LLM 边想边搜:只有 outcome reward(最终答对)会让搜索策略学不稳,只有 process reward(奖励每次搜索是否合理)又容易 over-fit 到局部步骤。作者把两者混合训练。

Qi Zhang,Shouqing Yang,Lirong Gao,Hao Chen,Xiaomeng Hu,Jinglei Chen ... 省略 1 位作者 ... ,Sheng Guo,Bo Zheng,Haobo Wang,Junbo Zhao
rlprocess-rewardsearch-agentDOIDBLP
6
泛读EMNLP 2025

Learning Like Humans: Advancing LLM Reasoning Capabilities via Adaptive Difficulty Curriculum Learning and Expert-Guided Self-Reformulation

Enci Zhang,Xingang Yan,Wei Lin,Tianxiang Zhang,Qianchun Lu
curriculumreasoningdifficultyDOIDBLP
6
泛读FindingsEMNLP 2025

SCoder: Progressive Self-Distillation for Bootstrapping Small-Scale Data Synthesizers to Empower Code LLMs

Xinyu Zhang,Changzhi Zhou,Linmei Hu,Luhao Zhang,Xiancai Chen,Haomin Fu,Yang Yang,Mengdi Zhang
self-distillationdata-synthesiscodeAnthologyDBLP
6
泛读EMNLP 2025

Dovetail: A CPU/GPU Heterogeneous Speculative Decoding for LLM inference

推测解码(speculative decoding)通常假设 draft model 和 target model 都在 GPU 上运行,但当 GPU 显存不足以同时容纳两个模型时效率大打折扣。Dovetail 要解决的是如何利用 CPU/GPU 异构计算来高效实现推测解码,让 CPU 跑 draft model、GPU 跑 target model,并通过流水线隐藏延迟。

Libo Zhang,Zhaoning Zhang,Xubaizhou,Rui Li,Zhiliang Tian,Songzhu Mei,Dongsheng Li
speculative-decodingheterogeneousinferenceDOIDBLP
6
泛读FindingsEMNLP 2025

DORM: Preference Data Weights Optimization for Reward Modeling in LLM Alignment

Reward model 训练中,不同偏好数据样本的质量差异很大,但现有方法对所有样本一视同仁地训练。DORM 要解决的是如何自动为偏好数据分配最优权重,使 reward model 更好地服务于下游对齐。

Rongzhi Zhang,Chenwei Zhang,Xinyang Zhang,Liang Qiu,Haoming Jiang,Yuchen Zhuang ... 省略 2 位作者 ... ,Xian Li,Bing Yin,Tuo Zhao,Chao Zhang
reward-modelpreference-dataalignmentAnthologyDBLP
6
泛读EMNLP 2025

Priority on High-Quality: Selecting Instruction Data via Consistency Verification of Noise Injection

指令微调数据的质量参差不齐,如何从大量候选数据中自动筛选高质量样本是 SFT 阶段的关键问题。现有方法多依赖外部模型打分或简单启发式,缺乏对数据内在一致性的利用。

Hong Zhang,Feng Zhao,Ruilin Zhao,Cheng Yan,Kangzheng Liu
data-selectioninstruction-tuningdata-qualityDOIDBLP
6
泛读EMNLP 2025

Large Language Models Badly Generalize across Option Length, Problem Types, and Irrelevant Noun Replacements

LLM 在选择题等结构化推理任务上的泛化能力很差——选项长度变化、问题类型切换、无关名词替换等表面变化就能显著影响模型表现。这揭示了 LLM 的推理可能依赖于表面模式而非真正的逻辑理解。

Guangxiang Zhao,Saier Hu,Xiaoqi Jian,Jinzhu Wu,Yuhan Wu,Lin Sun,Xiangzheng Zhang
generalizationrobustnessevaluationDOIDBLP
6
泛读EMNLP 2025

FISTAPruner: Layer-wise Post-training Pruning for Large Language Models

LLM 的后训练剪枝(post-training pruning)在高压缩比下性能下降严重。FISTAPruner 提出逐层剪枝方法来改善这一问题,核心思路是将剪枝问题分解为逐层的稀疏优化。

Pengxiang Zhao,Hanyu Hu,Ping Li,Yi Zheng,Zhefeng Wang,Xiaoming Yuan
pruningcompressionDOIDBLP
6
泛读EMNLP 2025

Probing and Boosting Large Language Models Capabilities via Attention Heads

LLM 的不同 attention head 承担不同功能,但如何系统地识别和利用这些功能性 head 来提升模型能力仍不清楚。本文通过探测 attention head 的功能来理解和增强 LLM 的能力。

Dezhi Zhao,Xin Liu,Xiaocheng Feng,Hui Wang,Bing Qin
attention-headsprobingDOIDBLP
6
泛读FindingsEMNLP 2025

Position IDs Matter: An Enhanced Position Layout for Efficient Context Compression in Large Language Models

上下文压缩(如 KV Cache 驱逐或 Prompt 压缩)会破坏原有 Token 的位置编码(Position IDs)连续性,导致模型丢失相对位置感知并引发性能断崖。过去的方法往往直接保留稀疏的原始 ID 或粗暴重新编号,未能解决注意力机制的扰动问题。

Runsong Zhao,Xin Liu,Xinyu Liu,Pengcheng Huang,Chunyang Xiao,Tong Xiao,JingBo Zhu
position-encodingcontext-compressionlong-contextAnthologyDBLP
6
泛读EMNLP 2025

Permutative Preference Alignment from Listwise Ranking of Human Judgments

现有的偏好对齐方法(如 DPO)主要依赖成对比较(Pairwise Comparison),未能充分利用人类标注中常见的列表级排序(Listwise Ranking)所包含的全局偏好密度信息。

Yang Zhao,Yixin Wang,Mingzhang Yin
preference-alignmentlistwisedpoDOIDBLP
6
泛读EMNLP 2025

When Long Helps Short: How Context Length in Supervised Fine-tuning Affects Behavior of Large Language Models

这篇论文要回答一个被低估但很实际的问题:SFT 时使用更长上下文,是否会改变模型在短上下文任务上的行为,而不仅仅提升长上下文能力。过去很多人默认‘长数据只影响长任务’,或者把上下文长度当作吞吐与成本问题处理,而没有系统分析它如何重塑模型的响应风格、注意分配和短程决策。

Yingming Zheng,Hanqi Li,Kai Yu,Lu Chen
sftlong-contexttraining-dynamicsDOIDBLP
6
泛读FindingsEMNLP 2025

Revisiting Pruning vs Quantization for Small Language Models

这篇论文要解决的是:对于小语言模型,剪枝和量化到底谁更有效,这个问题在实践里常被经验判断替代,但未必有系统结论。过去很多工作分别优化剪枝或量化,却少有在统一设置下比较二者在质量、延迟、显存和可部署性上的真实 trade-off。

Zihan Zhou,Simon Kurz,Zhixue Zhao
pruningquantizationsmall-lmAnthologyDBLP
6
泛读EMNLP 2025

QuZO: Quantized Zeroth-Order Fine-Tuning for Large Language Models

这篇论文要解决的是:当无法反传或显存/隐私受限时,零阶微调是一个可选方向,但它通常样本效率差、通信和存储成本高;如果再叠加量化,能否让 LLM 的 zero-order fine-tuning 真正可用。过去零阶方法在大模型上更多是概念验证,因为估计梯度本身就昂贵。

Jiajun Zhou,Yifan Yang,Kai Zhen,Ziyue Liu,Yequan Zhao,Ershad Banijamali,Athanasios Mouchtaris,Ngai Wong,Zheng Zhang
zeroth-orderquantizationfine-tuningDOIDBLP
6
泛读FindingsEMNLP 2025

DISCO Balances the Scales: Adaptive Domain- and Difficulty-Aware Reinforcement Learning on Imbalanced Data

Yuhang Zhou,Jing Zhu,Shengyi Qian,Zhuokai Zhao,Xiyao Wang,Xiaoyu Liu,Ming Li,Paiheng Xu,Wei Ai,Furong Huang
rldata-imbalanceadaptiveAnthologyDBLP
5
泛读IndustryEMNLP 2025

Reasoning-Enhanced Domain-Adaptive Pretraining of Multimodal Large Language Models for Short Video Content Governance

这篇论文要解决的是:把短视频内容治理从“每个违规类型各训一个小分类器”改成统一的多模态大模型判断,但现有 MLLM 的预训练分布和短视频平台数据差得太远,且违规定义往往依赖复杂规则与推理,直接拿来用效果不稳。过去做法主要靠任务拆分和重标注来补,代价高,而且跨问题泛化差;作者想用面向场景的 domain-adaptive pretraining 把感知、规则理解和推理一起补上。

Zixuan Wang,Yu Sun,Hongwei Wang,Baoyu Jing,Xiang Shen,Xin Dong,Zhuolin Hao,Hongyu Xiong,Yang Song
domain-adaptive-pretrainingmllmreasoningDOIarXivDBLP
5
泛读IndustryEMNLP 2025

Aligning LLMs for Multilingual Consistency in Enterprise Applications

这篇论文要解决的是:企业级多语言应用里,LLM 在非英语上的稳定性和一致性明显不足,即使接了 RAG 也会显著掉点,导致同一语义请求在不同语言上给出不同质量甚至不同结论。问题根源被作者归因为英语中心预训练和内部推理偏置;现有补救多依赖单语数据扩充或逐语种调参,成本高且难保跨语言对齐。

Amit Agarwal,Hansa Meghwani,Hitesh Laxmichand Patel,Tao Sheng,Sujith Ravi,Dan Roth
multilingualalignmentDOIarXivDBLP
5
泛读FindingsEMNLP 2025

EfficientXLang: Towards Improving Token Efficiency Through Cross-Lingual Reasoning

这篇论文要解决的是:英语是否真的是推理最省 token 的语言。现有 reasoning model 研究几乎默认用英语做 CoT 和评测,这让“token 效率”这个问题长期被忽略;但多语言预训练模型在不同语言上的分词长度、表达压缩率和推理习惯并不相同,英语未必最优。

Sanchit Ahuja,Praneetha Vaddamanu,Barun Patra
multilingualreasoningtoken-efficiencyAnthologyarXivDBLP
5
泛读FindingsEMNLP 2025

Are Knowledge and Reference in Multilingual Language Models Cross-Lingually Consistent?

这篇论文要解决的是:多语言模型在不同语言中是否真的保持一致的事实知识与指代理解。过去跨语言迁移通常看平均任务分数,但这无法回答同一知识在不同语言表达下是否被一致存取,更无法定位一致性在哪里断裂。作者因此用 code-mixed 的共指陈述来构造等价知识测试,并配合可解释分析看模型内部瓶颈。

Xi Ai,Mahardika Krisna Ihsani,Min-Yen Kan
multilingualconsistencyknowledgeAnthologyarXivDBLP
5
泛读EMNLP 2025

No Need for Explanations: LLMs can implicitly learn from mistakes in-context

在 few-shot 纠错提示中,错误答案是否必须配上详细的 rationale 才有用?作者质疑了这个默认假设。以前的做法(contrastive CoT、learn-from-mistakes 系列)都倾向于把错题和解释一起喂给模型,代价是数据贵、上下文长。

Lisa Alazraki,Maximilian Mozes,Jon Ander Campos,Yi Chern Tan,Marek Rei,Max Bartolo
Imperial College LondonCohereiclreasoningDOIarXivDBLP
5
泛读EMNLP 2025

CARMA: Enhanced Compositionality in LLMs via Advanced Regularisation and Mutual Information Alignment

LLM 的组合泛化仍然薄弱——见过组件 A 和 B,新组合 A+B 的输入就容易崩。以往的做法(架构改、finetune、数据增强)要么扩展性差要么收益递减。作者想做一种在 finetune 之上插入的轻量干预。

Nura Aljaafari,Danilo S. Carvalho,André Freitas
University of ManchesterIdiap Research Institutecompositionalityregularizationmutual-informationDOIarXivDBLP
5
泛读FindingsEMNLP 2025

Tag&Tab: Pretraining Data Detection in Large Language Models Using Keyword-Based Membership Inference Attack

判断某段文本是否出现在 LLM 的预训练数据中(membership inference)。现有方法多依赖 loss / perplexity 等整体分布指标,在大模型上信号弱、准确率有限。

Sagiv Antebi,Edan Habler,Asaf Shabtai,Yuval Elovici
Ben-Gurion University of the Negevmembership-inferencepretraining-datadata-detectionAnthologyDBLP
5
泛读EMNLP 2025

Improving Task Diversity in Label Efficient Supervised Finetuning of LLMs

SFT 的 label-efficient learning。以前的选样方法几乎都围绕 prompt-diversity 做(覆盖 prompt 空间),作者主张应该以 task-diversity 为核心维度:同一 task 内的 prompt 再多也带不来新信号。

Abhinav Arabelly,Jagrut Nemade,Robert D. Nowak,Jifan Zhang
University of Wisconsin-Madisonsftlabel-efficiencytask-diversityDOIarXivDBLP
5
泛读EMNLP 2025

Where Confabulation Lives: Latent Feature Discovery in LLMs

定位 LLM 生成中 confabulation(编造事实)在内部表示里的位置。以往研究多在 attention head 或 logit 层面找信号,作者转向 latent feature 维度。

Thibaud Ardoin,Yi Cai,Gerhard Wunder
Technische Universität Berlininterpretabilitylatent-featureshallucinationDOIDBLP
5
泛读FindingsEMNLP 2025

Can VLMs Recall Factual Associations From Visual References?

视觉语言模型(VLM)在多模态对齐中存在严重的表征割裂:给定文本实体能召回事实知识,但给定该实体的图像时却无法召回。

Dhananjay Ashok,Ashutosh Chaubey,Hirona Jacqueline Arai,Jonathan May,Jesse Thomason
vlmgroundingfactual-recallAnthologyarXivDBLP
5
泛读EMNLP 2025

How Do Large Vision-Language Models See Text in Image? Unveiling the Distinctive Role of OCR Heads

大型视觉语言模型(LVLM)如何从图像中提取文本(OCR)在内部机制上是一个黑盒,阻碍了对模型视觉文本理解能力的针对性优化。

Ingeol Baek,Hwan Chang,Sunghyun Ryu,Hwanhee Lee
vlminterpretabilityattention-headsDOIarXivDBLP
5
泛读FindingsEMNLP 2025

A Closer Look at Bias and Chain-of-Thought Faithfulness of Large (Vision) Language Models

虽然思维链(CoT)提升了大型视觉语言模型(LVLM)的性能,但模型输出的推理轨迹是否真实反映了其内部决策过程(即 CoT 忠实度)存疑,尤其是在面对视觉偏见时。

Sriram Balasubramanian,Samyadeep Basu,Soheil Feizi
cot-faithfulnessvlmbiasAnthologyarXivDBLP
5
泛读EMNLP 2025

GAP: a Global Adaptive Pruning Method for Large Language Models

(注:论文摘要缺失,基于标题推断)大型语言模型的剪枝通常依赖局部的逐层指标或静态的全局阈值,难以在压缩率和模型性能之间取得最优平衡。

Zhihua Ban,Haotian Ma,Siheng Zhang,Shengyu Liu,Xichen Chen,Ming Yang
pruningmodel-compressionefficiencyDOIDBLP
5
泛读FindingsEMNLP 2025

AfroXLMR-Social: Adapting Pre-trained Language Models for African Languages Social Media Text

多语言预训练模型在非洲语言上的预训练数据严重偏向宗教文本或维基百科,导致其在社交媒体等快速演变、充满噪音的下游领域表现糟糕。

Tadesse Destaw Belay,Israel Abebe Azime,Ibrahim Said Ahmad,David Ifeoluwa Adelani,Idris Abdulmumin,Abinew Ali Ayele,Shamsuddeen Hassan Muhammad,Seid Muhie Yimam
continual-pretrainingdomain-adaptationmultilingualAnthologyarXivDBLP
5
泛读IndustryEMNLP 2025

Learning to Translate Ambiguous Terminology by Preference Optimization on Post-Edits

在神经机器翻译(NMT)中,企业术语往往是多义的,正确的翻译取决于上下文和风格指南。传统的硬性词典约束或输入端软约束难以处理这种歧义。

Nathaniel Berger,Johannes Eschbach-Dymanus,Miriam Exel,Matthias Huck,Stefan Riezler
preference-optimizationmachine-translationalignmentDOIarXivDBLP
5
泛读FindingsEMNLP 2025

BanglaByT5: Byte-Level Modelling for Bangla

孟加拉语等形态丰富语言在传统BPE分词下面临词表膨胀和表示效率低下的问题。

Pramit Bhattacharyya,Arnab Bhattacharya
byte-level-lmtokenizer-freemultilingual-pretrainingAnthologyDBLP
5
泛读EMNLP 2025

so much depends / upon / a whitespace: Why Whitespace Matters for Poets and LLMs

预训练语料清洗和分词过程中对空白字符的压缩,破坏了诗歌等对空间排版高度敏感文本的结构语义。

Sriharsh Bhyravajjula,Melanie Walsh,Anna Preus,Maria Antoniak
tokenizerwhitespacegeneration-qualityDOIDBLP
5
泛读EMNLP 2025

Hallucination Detection in LLMs Using Spectral Features of Attention Maps

现有幻觉检测多依赖外部知识库,缺乏从模型内部注意力机制直接提取幻觉信号的轻量级方法。

Jakub Binkowski,Denis Janiak,Albert Sawczyn,Bogdan Gabrys,Tomasz Kajdanowicz
hallucination-detectionattention-mapsmechanistic-interpretabilityDOIDBLP
5
泛读EMNLP 2025

Efficient Compositional Multi-tasking for On-device Large Language Models

端侧LLM受限于显存,难以同时加载和并发处理多个针对不同任务微调的权重。

Ondrej Bohdal,Mete Ozay,Jijoong Moon,Kyeng-Hun Lee,Hyeonmok Ko,Umberto Michieli
Samsung AIon-devicemulti-taskingparameter-efficient-finetuningDOIDBLP
5
泛读FindingsEMNLP 2025

LLMs as a synthesis between symbolic and distributed approaches to language

理论界长期存在连接主义与符号主义的对立,需要解释LLM如何在分布式向量空间中实现类似符号操作的能力。

Gemma Boleda
Universitat Pompeu FabraICREAlanguage-modelingsymbolicrepresentationAnthologyDBLP
5
泛读EMNLP 2025

LiteraryQA: Towards Effective Evaluation of Long-document Narrative QA

现有长文本QA基准侧重信息检索,缺乏对长篇文学作品中复杂叙事结构和深层语义的有效评估。

Tommaso Bonomo,Luca Gioffrè,Roberto Navigli
Sapienza University of Romelong-contextbenchmarkevaluationDOIDBLP
5
泛读EMNLP 2025

CafGa: Customizing Feature Attributions to Explain Language Models

现有特征归因方法静态且通用,难以根据特定任务需求定制解释,导致在复杂LLM任务中解释性差。

Alan Boyle,Furui Cheng,Vilém Zouhar,Mennatallah El-Assady
ETH ZurichinterpretabilityattributionexplanationDOIDBLP
5
泛读FindingsEMNLP 2025

Acquiescence Bias in Large Language Models

评估和对齐阶段,LLM 倾向于无条件赞同用户的观点或预设(阿谀奉承/Acquiescence Bias),导致其作为裁判或助手的客观性受损。

Daniel Braun
biasbehaviorevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Enhancing Domain-Specific Encoder Models with LLM-Generated Data: How to Leverage Ontologies, and How to Do Without Them

领域专有 Encoder 模型缺乏高质量训练数据,而直接使用 LLM 合成数据时,如何有效注入领域本体(Ontology)知识仍缺乏系统性结论。

Marc Felix Brinner,Tarek Al Mustafa,Sina Zarrieß
synthetic-datadata-qualitydomain-adaptationAnthologyDBLP
5
泛读EMNLP 2025

Adaptively profiling models with task elicitation

静态 Benchmark 无法全面反映 LLM 的真实能力边界,需要一种自适应的方法来动态“诱导”并描绘模型能胜任的任务画像。

Davis Brown,Prithvi Balehannina,Helen Jin,Shreya Havaldar,Hamed Hassani,Eric Wong
evaluationelicitationcapabilityDOIDBLP
5
泛读EMNLP 2025

Enhancing Reasoning Abilities of Small LLMs with Cognitive Alignment

小参数量 LLM 的复杂推理能力存在明显瓶颈,传统的知识蒸馏难以有效传递大模型的内在推理逻辑。

Wenrui Cai,Chengyu Wang,Junbing Yan,Jun Huang,Xiangzhong Fang
Alibabareasoningsmall-llmalignmentDOIDBLP
5
泛读EMNLP 2025

DINT Transformer

Transformer 架构在处理长序列或大规模部署时计算开销过大,需要更底层的架构或数值表示优化。

Yueyang Cang,Yuhang Liu,Xiaoteng Zhang,Erlu Zhao,Li Shi
transformerarchitectureefficiencyDOIDBLP
5
泛读EMNLP 2025

Towards Advanced Mathematical Reasoning for LLMs via First-Order Logic Theorem Proving

LLM 在处理复杂数学推理时容易产生逻辑幻觉,自然语言的非形式化特性导致缺乏严格的验证机制。

Chuxue Cao,Mengze Li,Juntao Dai,Jinluan Yang,Zijian Zhao,Shengyu Zhang,Weijie Shi,Chengzhong Liu,Sirui Han,Yike Guo
reasoningmaththeorem-provingDOIDBLP
5
泛读FindingsEMNLP 2025

Perspective-driven Preference Optimization with Entropy Maximization for Diverse Argument Generation

在偏好优化(如 DPO/PPO)过程中,模型容易陷入模式崩溃,导致生成的论述缺乏多样性。

Yilin Cao,Ruike Zhang,Penghui Wei,Qingchao Kong,Wenji Mao
CASIApreference-optimizationalignmententropyAnthologyDBLP
5
泛读EMNLP 2025

HydraOpt: Navigating the Efficiency-Performance Trade-off of Adapter Merging

这篇工作要解决的是:adapter merging 常被当成低成本多任务适配的工程捷径,但它在性能和效率之间没有统一的优化准则,实际合并时常常要靠暴力试配方。作者把问题明确成一个 trade-off 导航问题:在给定预算下,怎么选合并策略,既不把推理/存储开销做高,也不把任务性能合并没了。

Taha Ceritli,Ondrej Bohdal,Mete Ozay,Jijoong Moon,Kyeng-Hun Lee,Hyeonmok Ko,Umberto Michieli
adaptersmergingefficiencyDOIDBLP
5
泛读FindingsEMNLP 2025

Assessing Socio-Cultural Alignment and Technical Safety of Sovereign LLMs

这篇工作要解决的是:所谓 sovereign LLM 往往被政策和市场需求推动,但缺少一套同时覆盖社会文化对齐与技术安全的评估框架,导致模型是否真的“适合本地部署”常被单一 benchmark 代替。作者把问题从一般能力评测转向“价值适配 + 安全边界”的双维度审视。

Kyubyung Chae,Gihoon Kim,Gyuseong Lee,Taesup Kim,Jaejin Lee,Heejin Kim
alignmentsafetyevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

LoRA-MGPO: Mitigating Double Descent in Low-Rank Adaptation via Momentum-Guided Perturbation Optimization

这篇工作要解决的是:LoRA 在低秩受限和数据有限的条件下也会出现 double descent 式的不稳定泛化,说明参数高效微调并不天然更稳。作者关注的不是再提一个更大的 rank,而是如何在低秩更新空间里抑制这种先好后坏再回升的训练行为。

Yupeng Chang,Chenlu Guo,Yi Chang,Yuan Wu
loraoptimizationdouble-descentAnthologyDBLP
5
泛读EMNLP 2025

Context-Aware Membership Inference Attacks against Pre-trained Large Language Models

这篇工作要解决的是:针对预训练 LLM 的 membership inference attack 往往忽略上下文,默认只要看目标样本本身的似然或暴露度就够了,但语言模型记忆和泄露常常强依赖提示、邻近 token 和语境匹配。作者因此提出 context-aware MIA,试图更真实地刻画训练样本是否能被推断出来。

Hongyan Chang,Ali Shahin Shamsabadi,Kleomenis Katevas,Hamed Haddadi,Reza Shokri
miamemorizationprivacyDOIDBLP
5
泛读FindingsEMNLP 2025

What Makes for Good Image Captions?

这篇工作要解决的是:图像 caption 的好坏长期被 CIDEr、BLEU 等自动指标代理,但这些指标和真实“好 caption”之间并不总一致,尤其在信息覆盖、细节准确性、风格自然度和下游用途之间常有冲突。作者直接问‘什么样的 caption 才算好’,说明目标不是再刷一个 caption 模型,而是重审评价标准本身。

Delong Chen,Samuel Cahyawijaya,Etsuko Ishii,Ho Shu Chan,Yejin Bang,Pascale Fung
captioningdata-qualityAnthologyDBLP
5
泛读FindingsEMNLP 2025

ReAlign: Structured Revision for Small Language Model Alignment

这篇工作要解决的是:小语言模型做 alignment 时容量更紧、容错更低,直接照搬大模型上的偏好优化或拒答训练,常会把有用能力一起损伤。作者提出 ReAlign,用 structured revision 的思路做小模型对齐,目标是在有限容量下把监督信号组织得更可学,而不是一味加更强偏好损失。

Ruijun Chen,Jiajian Guo,Hongzhan Chen,Fanqi Wan,Qifan Wang,Xiaojun Quan
alignmentslmAnthologyDBLP
5
泛读EMNLP 2025

ConsistentChat: Building Skeleton-Guided Consistent Multi-Turn Dialogues for Large Language Models from Scratch

Jiawei Chen,Xinyan Guan,Qianhao Yuan,Guozhao Mo,Weixiang Zhou,Yaojie Lu,Hongyu Lin,Ben He,Le Sun,Xianpei Han
dialoguesynthetic-dataDOIDBLP
5
泛读EMNLP 2025

Self-Augmented Preference Alignment for Sycophancy Reduction in LLMs

Chien Hung Chen,Hen-Hsen Huang,Hsin-Hsi Chen
sycophancypreference-alignmentDOIDBLP
5
泛读FindingsEMNLP 2025

CogAtom: From Cognitive Atoms to Olympiad-level Mathematical Reasoning in Large Language Models

Zhuofan Chen,Jiyuan He,Yichi Zhang,Xing Hu,Haoxing Wen,Jun Bai,Wenge Rong
math-reasoningdata-synthesisAnthologyDBLP
5
泛读EMNLP 2025

Sticker-TTS: Learn to Utilize Historical Experience with a Sticker-driven Test-Time Scaling Framework

Jie Chen,Jinhao Jiang,Yingqian Min,Zican Dong,Shijie Wang,Wayne Xin Zhao,Ji-Rong Wen
test-time-scalingreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

InteractSpeech: A Speech Dialogue Interaction Corpus for Spoken Dialogue Model

Yifu Chen,Shengpeng Ji,Ziqing Wang,Hanting Wang,Zhou Zhao
speech-lmdialoguecorpusAnthologyDBLP
5
泛读IndustryEMNLP 2025

Datasets and Recipes for Video Temporal Grounding via Reinforcement Learning

Ruizhe Chen,Tianze Luo,Zhiting Fan,Heqing Zou,Zhaopeng Feng,Guiyang Xie,Hansheng Zhang,Zhuochen Wang,Zuozhu Liu,Huaijian Zhang
rlvideotemporal-groundingDOIDBLP
5
泛读EMNLP 2025

MAgICoRe: Multi-Agent, Iterative, Coarse-to-Fine Refinement for Reasoning

Justin Chih-Yao Chen,Archiki Prasad,Swarnadeep Saha,Elias Stengel-Eskin,Mohit Bansal
multi-agentreasoningrefinementDOIDBLP
5
泛读FindingsEMNLP 2025

GuiLoMo: Allocating Experts and Ranks for LoRA-MoE via Bilevel Optimization with GuidedSelection Vectors

Xinrong Chen,Hengyuan Zhang,Yingmin Qiu,Xiao Liang,Ziyue Li,Guanyu Wang,Weiping Li,Tong Mo,Hayden Kwok-Hay So,Ngai Wong
loramoebilevelAnthologyDBLP
5
泛读FindingsEMNLP 2025

Chain-of-Thought Prompting Obscures Hallucination Cues in Large Language Models: An Empirical Evaluation

CoT prompting 在提升推理准确率的同时,可能掩盖了模型的幻觉线索——即 CoT 生成的中间步骤让幻觉更难被检测到。此前对 CoT 与幻觉之间关系的研究不够系统。

Jiahao Cheng,Tiancheng Su,Jia Yuan,Guoxiu He,Jiawei Liu,Xinqi Tao,Jingwen Xie,Huaxia Li
chain-of-thoughthallucinationpromptingAnthologyDBLP
5
泛读EMNLP 2025

DyePack: Provably Flagging Test Set Contamination in LLMs Using Backdoors

如何可证明地检测 LLM 是否在测试集上发生了数据污染(test set contamination)。现有检测方法多为统计启发式,缺乏可证明的保证。

Yize Cheng,Wenxiao Wang,Mazda Moayeri,Soheil Feizi
University of Marylanddata-contaminationbenchmarkbackdoorDOIDBLP
5
泛读FindingsEMNLP 2025

Accelerating LLM Reasoning via Early Rejection with Partial Reward Modeling

LLM 推理(如 CoT)过程中,很多推理路径在早期就已经走向错误,但现有方法仍会完整生成整条链。如何通过部分奖励模型(partial reward model)在早期拒绝低质量推理路径,从而加速推理。

Seyyed Saeid Cheshmi,Azal Ahmad Khan,Xinran Wang,Zirui Liu,Ali Anwar
reasoningearly-rejectionreward-modelAnthologyDBLP
5
泛读EMNLP 2025

M-Wanda: Improving One-Shot Pruning for Multilingual LLMs

现有的 one-shot 剪枝方法(如 Wanda)主要针对英语 LLM 设计和评估,在多语言模型上表现不佳。本文改进 Wanda 使其更适合多语言 LLM 的剪枝。

Rochelle Choenni,Ivan Titov
University of EdinburghpruningmultilingualcompressionDOIDBLP
5
泛读FindingsEMNLP 2025

Dub-S2ST: Textless Speech-to-Speech Translation for Seamless Dubbing

实现无文本中间表示的语音到语音翻译(textless S2ST),用于无缝配音(dubbing)场景。传统方法依赖 ASR→MT→TTS 级联,引入延迟和错误累积。

Jeongsoo Choi,Jaehun Kim,Joon Son Chung
KAISTspeech-to-speechdubbingtextlessAnthologyDBLP
5
泛读FindingsEMNLP 2025

CAC-CoT: Connector-Aware Compact Chain-of-Thought for Efficient Reasoning Data Synthesis Across Dual-System Cognitive Tasks

如何高效地合成 Chain-of-Thought 推理数据,特别是针对需要双系统认知(System 1 直觉 + System 2 推理)的任务。现有 CoT 数据合成方法往往冗长且不够紧凑。

Sunguk Choi,Yonghoon Kwon,Heondeuk Lee
chain-of-thoughtdata-synthesisreasoningAnthologyDBLP
5
泛读FindingsEMNLP 2025

Think Clearly: Improving Reasoning via Redundant Token Pruning

思维链(CoT)推理会生成大量冗余或无关的中间 token,这不仅增加了推理阶段的计算和显存开销,有时还会干扰模型自身的推理逻辑。

Daewon Choi,Jimin Lee,Jihoon Tack,Woomin Song,Saket Dingliwal,Sai Muralidhar Jayanthi,Bhavana Ganesh,Jinwoo Shin,Aram Galstyan,Sravan Babu Bodapati
token-pruningreasoninginference-efficiencyAnthologyDBLP
5
泛读EMNLP 2025

VoiceBBQ: Investigating Effect of Content and Acoustics in Social Bias of Spoken Language Model

语音语言模型(Spoken LM)不仅处理文本内容,还处理声学特征(如口音、语调),但现有的社会偏见基准(如 BBQ)仅针对纯文本,无法评估声学特征引发的偏见。

Junhyuk Choi,Ro-hoon Oh,Jihwan Seol,Bugeun Kim
speech-lmsocial-biasaudioDOIDBLP
5
泛读EMNLP 2025

Probing LLM World Models: Enhancing Guesstimation with Wisdom of Crowds Decoding

LLM 在面对需要物理直觉或数值常识的“费米估算”问题时表现挣扎,单次贪婪解码往往会产生严重的幻觉或数量级错误。

Yun-Shiuan Chuang,Sameer Narendran,Nikunj Harlalka,Alexander Cheung,Sizhe Gao,Siddharth Suresh,Junjie Hu,Timothy T. Rogers
world-modeldecodingprobingDOIDBLP
5
泛读EMNLP 2025

Where to show Demos in Your Prompt: A Positional Bias of In-Context Learning

在上下文学习(ICL)中,少样本示例(few-shot demos)的放置位置对模型最终的预测准确率有巨大影响,但这种位置偏置的具体规律缺乏系统性量化。

Kwesi A. Cobbina,Tianyi Zhou
iclpositional-biaspromptingDOIDBLP
5
泛读FindingsEMNLP 2025

InFact: Informativeness Alignment for Improved LLM Factuality

标准的预训练和 SFT 优化的是文本的流畅度和似然度,而 RLHF 往往迎合人类偏好(人类容易被自信但错误的详尽回答欺骗),这导致 LLM 频繁生成看似合理但缺乏事实依据的幻觉。

Roi Cohen,Russa Biswas,Gerard de Melo
factualityalignmentinformativenessAnthologyDBLP
5
泛读EMNLP 2025

RAGferee: Building Contextual Reward Models for Retrieval-Augmented Generation

现有的奖励模型(Reward Models)通常只评估最终回答的质量,而忽略了回答是否忠实于检索到的上下文。这使得用标准 RM 来优化 RAG 系统时,模型容易脱离上下文产生幻觉。

Andrei Catalin Coman,Ionut-Teodor Sorodoc,Leonardo F. R. Ribeiro,Bill Byrne,James Henderson,Adrià de Gispert
ragreward-modelcontextual-rewardDOIDBLP
5
泛读EMNLP 2025

Synthetic Data for Evaluation: Supporting LLM-as-a-Judge Workflows with EvalAssist

评估 LLM-as-a-judge(用大模型做裁判)的准确性需要大量高质量的元评估(Meta-evaluation)数据,而完全依赖人工标注这些数据既昂贵又难以扩展。

Martín Santillán Cooper,Zahra Ashktorab,Hyo Jin Do,Erik Miehling,Werner Geyer,Jasmina Gajcin,Elizabeth M. Daly,Qian Pan,Michael Desmond
synthetic-datallm-as-a-judgeevaluationDOIDBLP
5
泛读EMNLP 2025

Anecdoctoring: Automated Red-Teaming Across Language and Place

现有的红蓝对抗(Red-Teaming)高度以英语和西方文化为中心。简单地将英语越狱提示词翻译成其他语言,会遗漏特定文化和地域语境下独有的安全漏洞。

Alejandro Cuevas,Saloni Dash,Bharat Kumar Nayak,Dan Vann,Madeleine I. G. Daepp
red-teamingmultilingualsafetyDOIDBLP
5
泛读IndustryEMNLP 2025

Recover-LoRA: Data-Free Accuracy Recovery of Degraded Language Models via Low-Rank Adaptation

这篇论文要解决的是:模型在量化、剪枝、权重损伤或其他退化后,能否在完全没有原始训练数据的情况下恢复精度。过去的精度恢复通常依赖少量校准集、蒸馏样本或再训练数据,这在隐私受限、模型来源不明或部署端修复场景里并不总是可行。

Devleena Das,Rajeev Patwari,Ashish Sirasao
loramodel-recoveryparameter-efficientDOIDBLP
5
泛读FindingsEMNLP 2025

Ferret: Faster and Effective Automated Red Teaming with Reward-Based Scoring Technique

这篇工作关注自动化 red teaming 的两难:覆盖要广,但人工评估太慢;自动评估快,但往往噪声大、误报多。现有红队系统常在生成攻击样本和评估危险性之间失衡,因此很难既高效又可靠地扩展到大规模模型审计。

Pala Tej Deep,Vernon Toh,Rishabh Bhardwaj,Soujanya Poria
red-teamingreward-modelsafetyAnthologyDBLP
5
泛读EMNLP 2025

Identifying Unlearned Data in LLMs via Membership Inference Attacks

这篇论文要解决的是:如何识别 LLM 里哪些数据已经被成功“遗忘”,哪些其实还残留在模型中。现有 unlearning 评估经常依赖下游性能下降、模板攻击或少量案例检查,但这些信号要么太间接,要么覆盖不全;作者显然在尝试把 membership inference attack(成员推断攻击)变成 unlearned-data 识别工具。

Advit Deepak,Megan Mou,Jing Huang,Diyi Yang
unlearningmembership-inferencedata-privacyDOIDBLP
5
泛读EMNLP 2025

AgentPro: Enhancing LLM Agents with Automated Process Supervision

这篇论文要解决的是:LLM agent 的过程监督太贵,人工逐步标注不可扩展,导致 agent 常学到结果导向而不是过程可靠性。现有方法往往只监督最终成败或少量轨迹,无法系统覆盖检索、规划、工具调用这些中间步骤的错误来源。

Yuchen Deng,Shichen Fan,Naibo Wang,Xinkui Zhao,See-Kiong Ng
agentprocess-supervisionplanningDOIDBLP
5
泛读FindingsEMNLP 2025

MaskCD: Mitigating LVLM Hallucinations by Image Head Masked Contrastive Decoding

这篇论文解决的是 LVLM 幻觉问题,具体切入点是:如何在解码阶段更可靠地区分“来自图像的证据”和“来自语言先验的臆测”。许多现有方法要么改训练、要么做外部校验,部署成本较高;而纯解码式方法常因视觉信号太弱,无法稳定压制语言模型的惯性续写。

Jingyuan Deng,Yujiu Yang
contrastive-decodinghallucinationlvlmAnthologyDBLP
5
泛读FindingsEMNLP 2025

GUARD: Glocal Uncertainty-Aware Robust Decoding for Effective and Efficient Open-Ended Text Generation

开放式生成的解码策略在'质量-多样性-重复'三角上很难兼顾:greedy 和低温采样易重复,高温采样又容易跑飞。作者用不确定性指标做自适应解码。

Yuanhao Ding,Esteban Garces Arias,Meimingwei Li,Julian Rodemann,Matthias Aßenmacher,Danlu Chen,Gaojuan Fan,Christian Heumann,Chongsheng Zhang
decodinguncertaintytext-generationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Do LVLMs Know What They Know? A Systematic Study of Knowledge Boundary Perception in LVLMs

LVLM 是否知道自己不知道?纯文本 LLM 上的知识边界感知已经研究不少,但 LVLM 多了视觉通道,幻觉来源更复杂。作者做了系统性评测。

Zhikai Ding,Shiyu Ni,Keping Bi
lvlmcalibrationknowledgeAnthologyDBLP
5
泛读EMNLP 2025

Are Generative Models Underconfident? Better Quality Estimation with Boosted Model Probability

用生成模型自己的输出概率做 quality estimation(比如翻译质量评分)往往偏低、区分度不够。作者发现这是系统性的 underconfidence,并想办法修正。

Tu Anh Dinh,Jan Niehues
KITcalibrationprobabilityquality-estimationDOIDBLP
5
泛读EMNLP 2025

Scaling Rich Style-Prompted Text-to-Speech Datasets

风格可控的 TTS 数据太少、太窄——现有 style-prompted 数据集要么规模小,要么风格标签稀疏(只有情绪几类),导致 TTS 模型在自然语言风格指令下泛化差。

Anuj Diwan,Zhisheng Zheng,David Harwath,Eunsol Choi
UT AustinttsdatastyleDOIDBLP
5
泛读EMNLP 2025

"I've Decided to Leak": Probing Internals Behind Prompt Leakage Intents

System prompt 泄漏是部署中的现实风险,但一直被当作黑盒输入/输出层面的问题。这篇想从模型内部 representation 去看:当模型“决定泄漏”时,内部是否有可识别的信号。

Jianshuo Dong,Yutong Zhang,Liu Yan,Zhenyu Zhong,Tao Wei,Ke Xu,Minlie Huang,Chao Zhang,Han Qiu
prompt-leakageprobinginterpretabilityDOIDBLP
5
泛读EMNLP 2025

DSCD: Large Language Model Detoxification with Self-Constrained Decoding

去毒通常靠外部分类器或额外训练,代价是流畅度下降或需要额外模型;想要一种不依赖外部判别器、推理时就能做的去毒解码。

Ming Dong,Jinkui Zhang,Bolong Zheng,Xinhui Tu,Po Hu,Tingting He
detoxificationdecodingsafetyDOIDBLP
5
泛读EMNLP 2025

Reason to Rote: Rethinking Memorization in Reasoning

Yupei Du,Philipp Mondorf,Silvia Casola,Yuekun Yao,Robert Litschko,Barbara Plank
memorizationreasoninggeneralizationDOIDBLP
5
泛读EMNLP 2025

Correlation-Aware Example Selection for In-Context Learning with Nonsymmetric Determinantal Point Processes

Qiunan Du,Zhiliang Tian,Zhen Huang,Kailun Bian,Tianlun Liu,Zhaoning Zhang,Xinwang Liu,Feng Liu,Dong Sheng Li
iclexample-selectiondppDOIDBLP
5
泛读FindingsEMNLP 2025

Representation-based Broad Hallucination Detectors Fail to Generalize Out of Distribution

Zuzanna Dubanowska,Maciej Zelaszczyk,Michal Brzozowski,Paolo Mandica,Michal P. Karpowicz
hallucinationoodrepresentationAnthologyDBLP
5
泛读FindingsEMNLP 2025

How Sampling Affects the Detectability of Machine-written texts: A Comprehensive Study

Matthieu Dubois,François Yvon,Pablo Piantanida
samplingdetectionmachine-generated-textAnthologyDBLP
5
泛读FindingsEMNLP 2025

GeLoRA: Geometric Adaptive Ranks For Efficient LoRA Fine-tuning

Abdessalam Ed-dib,Zhanibek Datbayev,Amine Mohamed Aboussalah
lorapeftadaptive-rankAnthologyDBLP
5
泛读FindingsEMNLP 2025

Beyond Contrastive Learning: Synthetic Data Enables List-wise Training with Multiple Levels of Relevance

标准对比学习(如 InfoNCE)将文档相关性简化为二元(正/负)标签,无法建模真实检索场景中多层级的相关性梯度,导致模型对难负样本和部分相关文档的区分能力不足。

Reza Esfandiarpoor,George Zerveas,Ruochen Zhang,Macton Mgonzo,Carsten Eickhoff,Stephen H. Bach
Brown UniversityUniversity of Tübingensynthetic-dataretrievallistwiseAnthologyDBLP
5
泛读EMNLP 2025

Tool Preferences in Agentic LLMs are Unreliable

当前 Agent 评估默认 LLM 选择某个工具是因为它“理解”该工具最合适,但这种偏好是否真正基于语义逻辑,还是仅仅是对提示词格式的过拟合?

Kazem Faghih,Wenxiao Wang,Yize Cheng,Siddhant Bharti,Gaurang Sriramanan,Sriram Balasubramanian,Parsa Hosseini,Soheil Feizi
tool-useagentreliabilityDOIDBLP
5
泛读FindingsEMNLP 2025

Language-Specific Layer Matters: Efficient Multilingual Enhancement for Large Vision-Language Models

将大规模视觉语言模型(VLM)扩展到多语言时,全参微调成本极高且容易导致灾难性遗忘(破坏原有的英文图文对齐能力)。

Yuchun Fan,Yilin Wang,Yongyu Mu,Lei Huang,Bei Li,Xiaocheng Feng,Tong Xiao,JingBo Zhu
vlmmultilinguallayer-analysisAnthologyDBLP
5
泛读FindingsEMNLP 2025

End-to-End Optimization for Multimodal Retrieval-Augmented Generation via Reward Backpropagation

现有的多模态 RAG 系统通常将检索器和生成器作为两个独立的冻结模块,检索器无法根据生成器的最终输出质量来调整检索策略。

Zhiyuan Fan,Longfei Yun,Ming Yan,Yumeng Wang,Dadi Guo,Brian Mak,James T. Kwok,Yi R. Fung
Alibaba DAMO AcademyragmultimodalrewardAnthologyDBLP
5
泛读EMNLP 2025

VisiPruner: Decoding Discontinuous Cross-Modal Dynamics for Efficient Multimodal LLMs

高分辨率图像在 MLLM 中会产生海量视觉 token,导致严重的 KV-cache 膨胀和注意力计算瓶颈,而其中大部分背景 token 对文本生成毫无帮助。

Yingqi Fan,Anhao Zhao,Jinlan Fu,Junlong Tong,Hui Su,Yijie Pan,Wei Zhang,Xiaoyu Shen
TencentBaidumllmpruningtoken-reductionDOIDBLP
5
泛读FindingsEMNLP 2025

Towards More Efficient Post-training via Fourier Domain Adapter Framework

传统的 PEFT 方法(如 LoRA)在空间/权重域进行低秩更新,难以高效捕捉全局的、低频的语义模式,导致在复杂指令微调时表达能力受限。

Yijia Fan,Jusheng Zhang,Keze Wang
Sun Yat-sen UniversityadapterpeftfourierAnthologyDBLP
5
泛读EMNLP 2025

NL2Lean: Translating Natural Language into Lean 4 through Multi-Aspect Reinforcement Learning

将自然语言数学题转化为 Lean 4 形式化代码(Autoformalization)极度缺乏高质量平行语料,且依赖 SFT 难以满足编译器严苛的语法和类型要求。

Yue Fang,Shaohan Huang,Xin Yu,Haizhen Huang,Zihan Zhang,Weiwei Deng,Furu Wei,Feng Sun,Qi Zhang,Zhi Jin
Microsoft Researchformal-mathrlleanDOIDBLP
5
泛读FindingsEMNLP 2025

LastingBench: Defend Benchmarks Against Knowledge Leakage

这篇工作要解决的是:现有 LLM benchmark 很容易被预训练或后训练数据污染,导致评测分数越来越像记忆检测而不是能力测量。作者关注的不是再做一个更难题集,而是如何让 benchmark 在模型持续迭代、公开数据不断回流的环境里更抗知识泄漏,避免评测快速失效。

Yixiong Fang,Tianran Sun,Yuling Shi,Min Wang,Xiaodong Gu
benchmarkcontaminationevaluationAnthologyDBLP
5
泛读EMNLP 2025

AIMMerging: Adaptive Iterative Model Merging Using Training Trajectories for Language Model Continual Learning

这篇工作要解决的是:持续学习中的模型合并常常忽略训练轨迹,只拿几个终态 checkpoint 做静态融合,因此难以同时保住旧知识和新知识。作者试图用训练过程信息做自适应迭代合并,以缓解 continual learning 里的遗忘与冲突。

Yujie Feng,Jian Li,Xiaoyu Dong,Pengfei Xu,Xiaohui Zhou,Yujia Zhang ... 省略 1 位作者 ... ,Yasha Wang,Alan Zhao,Xu Chu,Xiao-Ming Wu
continual-learningmodel-mergingDOIDBLP
5
泛读FindingsEMNLP 2025

Enhancing Speech-to-Speech Dialogue Modeling with End-to-End Retrieval-Augmented Generation

这篇工作要解决的是:端到端 speech-to-speech dialogue 模型在长程对话记忆、事实一致性和风格延续上容易退化,作者尝试把检索增强直接接到语音到语音生成链路里。过去不少方法会先转文本再做 RAG,或者只在文本层加外部记忆,这会损失语音层信息并增加级联误差。

Pengchao Feng,Ziyang Ma,Wenxi Chen,Yao Li,Sheng Wang,Kai Yu,Xie Chen
speech-lmragAnthologyDBLP
5
泛读FindingsEMNLP 2025

MT-RewardTree: A Comprehensive Framework for Advancing LLM-Based Machine Translation via Reward Modeling

这篇工作要解决的是:LLM-based MT 的 reward modeling 仍然过于单一,单个分数难以同时覆盖忠实性、流畅性、术语、格式约束和细粒度错误类型,因此 RL 或 reranking 容易学偏。作者提出 RewardTree,目标是把翻译奖励从单值黑箱改成结构化框架。

Zhaopeng Feng,Jiahan Ren,Jiayuan Su,Jiamei Zheng,Hongwei Wang,Zuozhu Liu
reward-modeltranslationAnthologyDBLP
5
泛读FindingsEMNLP 2025

CoMoE: Contrastive Representation for Mixture-of-Experts in Parameter-Efficient Fine-tuning

这篇工作要解决的是:MoE 做参数高效微调时,专家路由和表示空间往往不够稳定,导致专家使用塌缩、任务区分不清或 PEFT 增益有限。作者提出用对比式表示来改进 MoE-PEFT,使专家分工更明确。

Jinyuan Feng,Chaopeng Wei,Tenghai Qiu,Tianyi Hu,Zhiqiang Pu
moepeftcontrastiveAnthologyDBLP
5
泛读EMNLP 2025

Enhancing RLHF with Human Gaze Modeling

Karim Galliamov,Ivan Titov,Ilya Pershin
rlhfhuman-feedbackDOIDBLP
5
泛读EMNLP 2025

Understanding LLMs' Cross-Lingual Context Retrieval: How Good It Is And Where It Comes From

Changjiang Gao,Hankun Lin,Xin Huang,Xue Han,Junlan Feng,Chao Deng,Jiajun Chen,Shujian Huang
multilingualcontext-retrievalanalysisDOIDBLP
5
泛读FindingsEMNLP 2025

DeCoRe: Decoding by Contrasting Retrieval Heads to Mitigate Hallucinations

Aryo Pradipta Gema,Chen Jin,Ahmed Abdulaal,Tom Diethe,Philip Alexander Teare,Beatrice Alex,Pasquale Minervini,Amrutha Saseendran
hallucinationretrieval-headscontrastive-decodingAnthologyDBLP
5
泛读FindingsEMNLP 2025

A Survey of Multilingual Reasoning in Language Models

Akash Ghosh,Debayan Datta,Sriparna Saha,Chirag Agarwal
multilingualreasoningsurveyAnthologyDBLP
5
泛读EMNLP 2025

Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts

Michal Golovanevsky,William Rudman,Michael A. Lepori,Amir Bar,Ritambhara Singh,Carsten Eickhoff
vlmknowledge-priorscounterfactualDOIDBLP
5
泛读FindingsEMNLP 2025

CARVQ: Corrective Adaptor with Group Residual Vector Quantization for LLM Embedding Compression

Dayin Gou,Sanghyun Byun,Nilesh Malpeddi,Gabrielle De Micheli,Prathamesh Vaste,Jacob Song,Woo Seong Chung
embedding-compressionvector-quantizationkv-cacheAnthologyDBLP
5
泛读EMNLP 2025

Corrupted but Not Broken: Understanding and Mitigating the Negative Impacts of Corrupted Data in Visual Instruction Tuning

Yunhao Gou,Hansi Yang,Zhili Liu,Kai Chen,Yihan Zeng,Lanqing Hong ... 省略 1 位作者 ... ,Qun Liu,Bo Han,James Kwok,Yu Zhang
data-qualityvisual-instruction-tuningcorrupted-dataDOIDBLP
5
泛读FindingsEMNLP 2025

Temporal Alignment of Time Sensitive Facts with Activation Engineering

Sanjay Govindan,Maurice Pagnucco,Yang Song
activation-engineeringtemporal-knowledgerepresentation-editingAnthologyDBLP
5
泛读EMNLP 2025

zFLoRA: Zero-Latency Fused Low-Rank Adapters

Dhananjaya Gowda,Seoha Song,Harshith Goka,Junhyun Lee
loraadapter-fusioninference-efficiencyDOIDBLP
5
泛读EMNLP 2025

Breaking Bad Tokens: Detoxification of LLMs Using Sparse Autoencoders

Agam Goyal,Vedant Rathi,William Yeh,Yian Wang,Yuen Chen,Hari Sundaram
sparse-autoencoderdetoxificationinterpretabilityDOIDBLP
5
泛读EMNLP 2025

Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers

Tommaso Green,Martin Gubri,Haritz Puerto,Sangdoo Yun,Seong Joon Oh
reasoning-modelprivacychain-of-thoughtDOIDBLP
5
泛读FindingsEMNLP 2025

NLoRA: Nyström-Initiated Low-Rank Adaptation for Large Language Models

Chenlu Guo,Yi Chang,Yuan Wu
lorapeftoptimizationAnthologyDBLP
5
泛读EMNLP 2025

ToM: Leveraging Tree-oriented MapReduce for Long-Context Reasoning in Large Language Models

Jiani Guo,Zuchao Li,Jie Wu,Qianren Wang,Yun Li,Lefei Zhang,Hai Zhao,Yu-Jiu Yang
long-contextreasoningmapreduceDOIDBLP
5
泛读EMNLP 2025

CARE: Multilingual Human Preference Learning for Cultural Awareness

Geyang Guo,Tarek Naous,Hiromi Wakaki,Yukiko Nishimura,Yuki Mitsufuji,Alan Ritter,Wei Xu
preference-learningmultilingualalignmentDOIDBLP
5
泛读EMNLP 2025

Precise In-Parameter Concept Erasure in Large Language Models

解决 LLM 中特定概念(如隐私、有害知识、版权内容)难以在参数层面被精准擦除,且容易损害模型通用能力的问题。

Yoav Gur-Arieh,Clara Suslik,Yihuai Hong,Fazl Barez,Mor Geva
unlearningconcept-erasurealignmentDOIDBLP
5
泛读FindingsEMNLP 2025

Rating Roulette: Self-Inconsistency in LLM-As-A-Judge Frameworks

揭示并量化 LLM-as-a-Judge 框架中存在的严重自我不一致性(Self-Inconsistency)问题。

Rajarshi Haldar,Julia Hockenmaier
UIUCllm-as-a-judgeevaluationreward-modelAnthologyDBLP
5
泛读FindingsEMNLP 2025

SafeSwitch: Steering Unsafe LLM Behavior via Internal Activation Signals

解决传统基于外部审查或输出过滤的 LLM 安全机制延迟高、且容易被复杂越狱攻击绕过的问题。

Peixuan Han,Cheng Qian,Xiusi Chen,Yuji Zhang,Heng Ji,Denghui Zhang
UIUCsafetyactivation-steeringrepresentationAnthologyDBLP
5
泛读EMNLP 2025

Towards a Unified Paradigm of Concept Editing in Large Language Models

解决当前 LLM 概念编辑(Concept Editing)领域方法碎片化、缺乏统一评估和建模框架的问题。

Zhuowen Han,Xinwei Wu,Dan Shi,Renren Jin,Deyi Xiong
Tianjin Universitymodel-editingalignmentknowledge-editingDOIDBLP
5
泛读EMNLP 2025

CAT: Causal Attention Tuning For Injecting Fine-grained Causal Knowledge into Large Language Models

解决 LLM 难以捕捉细粒度因果关系,导致在复杂逻辑推理任务中表现不佳的问题。

Kairong Han,Wenshuo Zhao,Ziyu Zhao,Ye Jun Jian,Lujia Pan,Kun Kuang
attentionknowledge-injectionfine-tuningDOIDBLP
5
泛读EMNLP 2025

From Understanding to Generation: An Efficient Shortcut for Evaluating Language Models

这篇论文要解决的是:能否用更便宜的“理解型”评测去替代昂贵的“生成型”评测,从而快速估计语言模型的真实生成能力。过去社区大量依赖开放式生成评测,但这类评测成本高、方差大、还容易被解码设置污染,因此值得重新寻找更稳定的代理指标。

Viktor Hangya,Fabian Küch,Darina Gold
evaluationgenerationefficiencyDOIDBLP
5
泛读FindingsEMNLP 2025

When Instructions Multiply: Measuring and Estimating LLM Capabilities of Multiple Instructions Following

这篇论文关注一个被单指令 benchmark 掩盖的问题:LLM 在面对多个同时成立的指令时,真实能力到底有多强,以及能否被可靠估计。过去很多指令跟随评测默认只有单一目标,因而高估了模型在复杂用户约束下的可用性。

Keno Harada,Yudai Yamazaki,Masachika Taniguchi,Edison Marrese-Taylor,Takeshi Kojima,Yusuke Iwasawa,Yutaka Matsuo
instruction-followingevaluationcapabilityAnthologyDBLP
5
泛读FindingsEMNLP 2025

Decoding Uncertainty: The Impact of Decoding Strategies for Uncertainty Estimation in Large Language Models

这篇论文研究的是:语言模型不确定性估计到底有多大程度取决于 decoding,而不仅仅取决于模型本身。过去很多工作把采样、温度、beam search 当成后处理细节,但如果不确定性结论对解码非常敏感,那么当前大量风险评估可能并不稳固。

Wataru Hashimoto,Hidetaka Kamigaito,Taro Watanabe
decodinguncertaintysamplingAnthologyDBLP
5
泛读EMNLP 2025

LogicTree: Structured Proof Exploration for Coherent and Rigorous Logical Reasoning with Large Language Models

这篇论文要解决的是:LLM 做逻辑推理时,单链式生成容易前后不一致、遗漏分支,也缺乏严格证明结构。现有 CoT 往往擅长给出一条看似顺畅的路径,但不擅长系统探索可能的证明空间。

Kang He,Kaushik Roy
reasoningsearchproofDOIDBLP
5
泛读EMNLP 2025

Layer-wise Minimal Pair Probing Reveals Contextual Grammatical-Conceptual Hierarchy in Speech Representations

这篇工作关注的是:语音表征中的语法信息和概念信息是否按层次、按上下文逐步形成,而不是混在一起被一次性编码。过去对 speech representation 的分析常把语音当成单一表征黑箱,只看最终任务准确率,导致我们很难判断模型先学到的是音系、语法还是语义概念。

Linyang He,Qiaolin Wang,Xilin Jiang,Nima Mesgarani
speechrepresentationprobingDOIDBLP
5
泛读FindingsEMNLP 2025

Distill Visual Chart Reasoning Ability from LLMs to MLLMs

这篇工作要解决的是:通用 MLLM 在图表推理上常常看得见图却读不准逻辑,而纯文本 LLM 已经积累了更强的结构化推理能力,如何把后者蒸馏到前者。以往图表能力提升往往依赖额外标注数据或更强视觉编码器,但这两条路都昂贵,而且不一定真正补上推理短板。

Wei He,Zhiheng Xi,Wanxu Zhao,Xiaoran Fan,Yiwen Ding,Zifei Shan,Tao Gui,Qi Zhang,Xuanjing Huang
mllmdistillationchart-reasoningAnthologyDBLP
5
泛读FindingsEMNLP 2025

Guaranteed Guess: A Language Modeling Approach for CISC-to-RISC Transpilation with Testing Guarantees

这篇工作要解决的是:把 CISC 指令转成 RISC 指令时,纯程序综合或纯规则系统覆盖不足,而纯 LM 生成又缺少正确性保证,如何把语言建模和可验证测试结合起来。这个问题虽然偏程序语言,但其核心仍是 sequence modeling 在强约束空间里的可靠生成。

Ahmed Heakl,Sarim Hashmi,Chaimaa Abi,Celine Lee,Abdulrahman Mahmoud
language-modelingcodetranspilationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Distilling Many-Shot In-Context Learning into a Cheat Sheet

这篇工作关注的是:many-shot ICL 虽然强,但推理时要塞很多示例,成本高且上下文脆弱,能否把这种 in-context 行为蒸馏成一张更短、更稳定的“cheat sheet”。过去常见做法是直接减少 exemplars 或做 retrieval,但这通常牺牲性能,因为模型真正利用的不是单个示例,而是示例集合中归纳出的任务规则。

Ukyo Honda,Soichiro Murakami,Peinan Zhang
iclmany-shotdistillationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Measuring Sycophancy of Language Models in Multi-turn Dialogues

这篇工作要解决的是:现有 sycophancy 评测多是单轮问答,无法反映真实对话里模型如何在多轮互动中逐步迎合用户。这个问题现在值得重做,因为部署环境里大多数对齐失败不是单轮发生,而是在用户持续施压、暗示立场或暴露偏好后累积出现。

Jiseung Hong,Grace Byun,Seungone Kim,Kai Shu
sycophancyalignmentevaluationAnthologyDBLP
5
泛读EMNLP 2025

Playpen: An Environment for Exploring Learning From Dialogue Game Feedback

Nicola Horst,Davide Mazzaccara,Antonia Schmidt,Michael Sullivan,Filippo Momentè,Luca Franceschetti ... 省略 6 位作者 ... ,Oliver Lemon,David Schlangen,Mario Giulianelli,Alessandro Suglia
rldialogueenvironmentDOIDBLP
5
泛读FindingsEMNLP 2025

WebCoT: Enhancing Web Agent Reasoning by Reconstructing Chain-of-Thought in Reflection, Branching, and Rollback

Minda Hu,Tianqing Fang,Jianshu Zhang,Jun-Yu Ma,Zhisong Zhang,Jingyan Zhou,Hongming Zhang,Haitao Mi,Dong Yu,Irwin King
web-agentcotreasoningAnthologyDBLP
5
泛读EMNLP 2025

Context Reasoner: Incentivizing Reasoning Capability for Contextualized Privacy and Safety Compliance via Reinforcement Learning

Wenbin Hu,Haoran Li,Huihao Jing,Qi Hu,Ziqian Zeng,Sirui Han,Heli Xu,Tianshu Chu,Peizhao Hu,Yangqiu Song
rlreasoningsafetyDOIDBLP
5
泛读EMNLP 2025

Quantifying Language Disparities in Multilingual Large Language Models

Songbo Hu,Ivan Vulic,Anna Korhonen
multilingualevaluationDOIDBLP
5
泛读FindingsEMNLP 2025

No Free Lunch: Retrieval-Augmented Generation Undermines Fairness in LLMs, Even for Vigilant Users

Mengxuan Hu,Hongyi Wu,Ronghang Zhu,Zihan Guan,Dongliang Guo,Daiqing Qi,Sheng Li
ragfairnessAnthologyDBLP
5
泛读FindingsEMNLP 2025

Evaluating Test-Time Scaling LLMs for Legal Reasoning: OpenAI o1, DeepSeek-R1, and Beyond

Yinghao Hu,Yaoyao Yu,Leilei Gan,Bin Wei,Kun Kuang,Fei Wu
test-time-scalingreasoningevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

RouterEval: A Comprehensive Benchmark for Routing LLMs to Explore Model-level Scaling Up in LLMs

Zhongzhan Huang,Guoming Ling,Yupei Lin,Yandong Chen,Shanshan Zhong,Hefeng Wu,Liang Lin
routingscalingbenchmarkAnthologyDBLP
5
泛读FindingsEMNLP 2025

LoRA-PAR: A Flexible Dual-System LoRA Partitioning Approach to Efficient LLM Fine-Tuning

Yining Huang,Bin Li,Keke Tang,Meilian Chen
5
泛读FindingsEMNLP 2025

BenchMAX: A Comprehensive Multilingual Evaluation Suite for Large Language Models

Xu Huang,Wenhao Zhu,Hanxu Hu,Conghui He,Lei Li,Shujian Huang,Fei Yuan
multilingualbenchmarkevaluationAnthologyDBLP
5
泛读IndustryEMNLP 2025

Auto-Weighted Group Relative Preference Optimization for Multi-Objective Text Generation Tasks

Yuki Ichihara,Yuu Jinnai
preference-optimizationmulti-objectivedpoDOIDBLP
5
泛读FindingsEMNLP 2025

AdaptMerge: Inference Time Adaptive Visual and Language-Guided Token Merging for Efficient Large Multimodal Models

Zahidul Islam,Mrigank Rochan
token-merginginference-efficiencyvlmAnthologyDBLP
5
泛读FindingsEMNLP 2025

Creative Preference Optimization

Mete Ismayilzada,Antonio Laverghetta Jr.,Simone Luchini,Reet Patel,Antoine Bosselut,Lonneke van der Plas,Roger E. Beaty
preference-optimizationcreativityalignmentAnthologyDBLP
5
泛读FindingsEMNLP 2025

Data Doping or True Intelligence? Evaluating the Transferability of Injected Knowledge in LLMs

探讨预训练或SFT阶段注入的知识,是真正被模型理解并能用于下游推理,还是仅仅被浅层记忆。

Essa Jan,Moiz Ali,Muhammad Saram Hassan,Muhammad Fareed Zaffar,Yasir Zaki
NYU Abu DhabiLUMSknowledge-injectiontransferabilitydata-contaminationAnthologyDBLP
5
泛读EMNLP 2025

PRISM: Efficient Long-Range Reasoning With Short-Context LLMs

如何在不扩展预训练上下文窗口、不增加KV Cache开销的情况下,让短上下文模型具备长文本推理能力。

Dulhan Jayalath,James Bradley Wendt,Nicholas Monath,Sandeep Tata,Beliz Gunel
Google Researchlong-contextreasoningshort-context-llmDOIDBLP
5
泛读EMNLP 2025

Harmful Prompt Laundering: Jailbreaking LLMs with Abductive Styles and Symbolic Encoding

揭示大模型安全对齐机制在面对“溯因推理风格”和“符号编码”组合时的脆弱性。

Seongho Joo,Hyukhun Koh,Kyomin Jung
Seoul National UniversityjailbreakalignmentsafetyDOIDBLP
5
泛读EMNLP 2025

SABER: Uncovering Vulnerabilities in Safety Alignment via Cross-Layer Residual Connection

揭示跨层残差连接(Residual Connections)如何成为绕过大模型安全对齐的漏洞。

Maithili Joshi,Palash Nandi,Tanmoy Chakraborty
IIT Delhialignmentsafetyresidual-streamDOIDBLP
5
泛读FindingsEMNLP 2025

LC-Eval: A Bilingual Multi-Task Evaluation Benchmark for Long-Context Understanding

解决现有长上下文评估基准多局限于英语、且容易被简单的检索机制破解的问题。

Sheikh Jubair,Arwa Omayrah,Amal Alshammari,Alhanoof Althnian,Abdulhamed Alothaimen,Norah A. Alzahrani,Shahad D. Alzaidi,Nora Al-Twairesh,Abdulmohsen Al-Thubaity
long-contextbenchmarkbilingualAnthologyDBLP
5
泛读FindingsEMNLP 2025

Can LLMs Truly Plan? A Comprehensive Evaluation of Planning Capabilities

剥离预训练数据记忆的干扰,严格评估大语言模型是否真正具备规划(Planning)能力。

Gayeon Jung,HyeonSeok Lim,Minjun Kim,Joon-ho Lim,KyungTae Lim,Hansaem Kim
planningreasoningbenchmarkAnthologyDBLP
5
泛读FindingsEMNLP 2025

Measuring Lexical Diversity of Synthetic Data Generated through Fine-Grained Persona Prompting

这篇工作关注的是:用细粒度 persona prompting 生成的合成数据,词汇多样性到底有多高,是否真的比普通提示更丰富。当前合成数据讨论常把规模和任务覆盖放在前面,而忽略 lexical diversity;这会导致数据看起来多,实际上措辞高度重复,训练增益很快饱和。

Gauri Kambhatla,Chantal Shaib,Venkata Govindarajan
synthetic-datapersonadata-qualityAnthologyDBLP
5
泛读FindingsEMNLP 2025

When Format Changes Meaning: Investigating Semantic Inconsistency of Large Language Models

这篇工作研究的是:当内容不变、只改表达格式时,LLM 是否会给出语义不一致的判断或回答。现有鲁棒性工作常测 paraphrase 或 prompt wording,但“格式变化”这个因素经常被低估;如果模型把列表、表格、JSON、自然段等形式当成不同语义线索,就会出现不该有的 meaning shift。

Cheongwoong Kang,Jongeun Baek,Yeonjea Kim,Jaesik Choi
format-sensitivitygeneralizationreasoningAnthologyDBLP
5
泛读FindingsEMNLP 2025

SwiftPrune: Hessian-Free Weight Pruning for Large Language Models

这篇工作要解决的是:LLM 权重剪枝里,二阶信息通常有效但代价太高,如何在不显式构造 Hessian 的情况下做高效剪枝。现有高质量 pruning 方法常依赖 Hessian 或其近似,计算和内存都重,不适合大模型快速压缩;作者提出 Hessian-free 的 SwiftPrune,目标是把二阶启发保留下来,同时去掉昂贵计算。

Yuhan Kang,Yang Shi,Mei Wen,Jun He,Jianchao Yang,Zeyu Xue,Jing Feng,Xinwang Liu
pruningefficiencycompressionAnthologyDBLP
5
泛读EMNLP 2025

Calibrating LLM Confidence by Probing Perturbed Representation Stability

Reza Khanmohammadi,Erfan Miahi,Mehrsa Mardikoraem,Simerjot Kaur,Ivan Brugere,Charese Smiley,Kundan Thind,Mohammad M. Ghassemi
calibrationconfidenceprobingDOIDBLP
5
泛读EMNLP 2025

Detecting LLM Hallucination Through Layer-wise Information Deficiency: Analysis of Ambiguous Prompts and Unanswerable Questions

Hazel Kim,Tom A. Lamb,Adel Bibi,Philip Torr,Yarin Gal
hallucinationlayer-analysisDOIDBLP
5
泛读FindingsEMNLP 2025

Drift: Decoding-time Personalized Alignments with Implicit User Preferences

Minbeom Kim,Kang-il Lee,Seongho Joo,Hwaran Lee,Thibaut Thonet,Kyomin Jung
personalizationdecodingalignmentAnthologyDBLP
5
泛读EMNLP 2025

Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens

Sohee Kim,Soohyun Ryu,Joonhyung Park,Eunho Yang
vlmhallucinationanalysisDOIDBLP
5
泛读EMNLP 2025

Memorization or Reasoning? Exploring the Idiom Understanding of LLMs

Jisu Kim,Youngwoo Shin,Uiji Hwang,Jihun Choi,Richeng Xuan,Taeuk Kim
memorizationreasoninganalysisDOIDBLP
5
泛读EMNLP 2025

CCQA: Generating Question from Solution Can Improve Inference-Time Reasoning in SLMs

Jinyoung Kim,Ji Won Yoon
reasoningslminference-timeDOIDBLP
5
泛读EMNLP 2025

Steering Language Models in Multi-Token Generation: A Case Study on Tense and Aspect

Alina Klerings,Jannik Brinkmann,Daniel Ruffinelli,Simone Paolo Ponzetto
steeringactivationanalysisDOIDBLP
5
泛读FindingsEMNLP 2025

TransBERT: A Framework for Synthetic Translation in Domain-Specific Language Modeling

Julien Knafou,Luc Mottin,Anaïs Mottaz,Alexandre Flament,Patrick Ruch
continual-pretraindomain-specificsynthetic-dataAnthologyDBLP
5
泛读EMNLP 2025

VideoPASTA: 7K Preference Pairs That Matter for Video-LLM Alignment

Yogesh Kulkarni,Pooyan Fazli
video-llmdpoalignmentDOIDBLP
5
泛读FindingsEMNLP 2025

Extracting Conceptual Spaces from LLMs Using Prototype Embeddings

Nitesh Kumar,Usashi Chatterjee,Steven Schockaert
embeddingsconceptual-spaceanalysisAnthologyDBLP
5
泛读EMNLP 2025

Conflicting Needles in a Haystack: How LLMs behave when faced with contradictory information

Murathan Kurfali,Robert Östling
long-contextanalysisconflictDOIDBLP
5
泛读FindingsEMNLP 2025

ToolHaystack: Stress-Testing Tool-Augmented Language Models in Realistic Long-Term Interactions

Beong-woo Kwak,Minju Kim,Dongha Lim,Hyungjoo Chae,Dongjin Kang,Sunghwan Kim,Dongil Yang,Jinyoung Yeo
tool-uselong-contextbenchmarkAnthologyDBLP
5
泛读FindingsEMNLP 2025

LLaVE: Large Language and Vision Embedding Models with Hardness-Weighted Contrastive Learning

Zhibin Lan,Liqiang Niu,Fandong Meng,Jie Zhou,Jinsong Su
embeddingcontrastive-learningvlmAnthologyDBLP
5
泛读FindingsEMNLP 2025

Training Language Models to Critique With Multi-agent Feedback

现有的 LLM 批评(Critique)模型通常依赖单一人类反馈或单一模型生成,难以覆盖多维度的缺陷,导致自我修正(Self-correction)的上限受限。

Tian Lan,Wenwei Zhang,Chengqi Lyu,Shuaibin Li,Chen Xu,Heyan Huang,Dahua Lin,Xian-Ling Mao,Kai Chen
Shanghai AI Labcritiquemulti-agentfeedbackAnthologyDBLP
5
泛读EMNLP 2025

Interpretation Meets Safety: A Survey on Interpretation Methods and Tools for Improving LLM Safety

LLM 安全性研究(如越狱防御)与可解释性研究(如特征归因)通常脱节,缺乏系统性利用内部表示来理解和提升模型安全性的框架。

Seongmin Lee,Aeree Cho,Grace C. Kim,Shengyun Peng,Mansi Phute,Duen Horng Chau
interpretabilitysafetysurveyDOIDBLP
5
泛读FindingsEMNLP 2025

Evaluating Step-by-step Reasoning Traces: A Survey

评估 CoT 等中间推理步骤的质量缺乏统一标准,通常只看最终答案对错(Outcome-based),忽视了推理过程的逻辑一致性和事实正确性。

Jinu Lee,Julia Hockenmaier
UIUCcotreasoningevaluationAnthologyDBLP
5
泛读EMNLP 2025

Does Localization Inform Unlearning? A Rigorous Examination of Local Parameter Attribution for Knowledge Unlearning in Language Models

知识编辑和遗忘领域普遍假设“找到存储知识的参数位置(Localization),然后修改它就能实现遗忘”,但这一假设的实际有效性存疑。

Hwiyeong Lee,Uiji Hwang,Hyelim Lim,Taeuk Kim
unlearninglocalizationknowledge-editingDOIDBLP
5
泛读FindingsEMNLP 2025

Efficient Latent Semantic Clustering for Scaling Test-Time Computation of LLMs

Test-time scaling(如 Self-Consistency)需要生成大量样本并进行多数投票,计算成本极高,且在文本空间容易被表面表达不同但语义相同的答案干扰。

Sungjae Lee,Hoyoung Kim,Jeongyeon Hwang,Eunhyeok Park,Jungseul Ok
test-time-computeclusteringinference-scalingAnthologyDBLP
5
泛读FindingsEMNLP 2025

Mitigating Attention Localization in Small Scale: Self-Attention Refinement via One-step Belief Propagation

小规模模型(Small Language Models)中注意力机制容易出现过度局部化(Attention Localization,即注意力坍缩到少数无意义 token 上),导致上下文利用率和推理能力下降。

Nakyung Lee,Yeongoon Kim,Minhae Oh,Suhwan Kim,Jin Woo Koo,Hyewon Jo,Jungwoo Lee
attentionbelief-propagationsmall-scaleAnthologyDBLP
5
泛读EMNLP 2025

AMQ: Enabling AutoML for Mixed-precision Weight-Only Quantization of Large Language Models

LLM 的混合精度量化(不同层使用不同 bit 位)搜索空间巨大,手动设计规则难以在给定的显存/延迟约束下达到精度与压缩率的最优平衡。

Sangjun Lee,Seung-taek Woo,Jungyu Jin,Changhun Lee,Eunhyeok Park
POSTECHquantizationautomlmixed-precisionDOIDBLP
5
泛读IndustryEMNLP 2025

DecEx-RAG: Boosting Agentic Retrieval-Augmented Generation with Decision and Execution Optimization via Process Supervision

Agentic RAG 在多步检索和生成中,决策(何时检索、检索什么)和执行(如何利用检索结果)容易出现级联错误,仅靠最终答案的反馈难以纠正中间步骤的偏差。

Yongqi Leng,Yikun Lei,Xikai Liu,Meizhi Zhong,Bojian Xiong,Yurong Zhang,Yan Gao,Yi Wu,Yao Hu,Deyi Xiong
ragprocess-supervisionagentDOIDBLP
5
泛读FindingsEMNLP 2025

Jailbreak Attack Initializations as Extractors of Compliance Directions

越狱攻击(Jailbreak)通常被视为输入空间的对抗扰动,但其在模型隐空间中究竟激活了什么特征、为何能绕过安全对齐尚不完全明确。

Amit Levi,Rom Himelstein,Yaniv Nemcovsky,Avi Mendelson,Chaim Baskin
jailbreaksafetymechanistic-interpretabilityAnthologyDBLP
5
泛读FindingsEMNLP 2025

More Documents, Same Length: Isolating the Challenge of Multiple Documents in RAG

RAG 系统在处理多文档时性能下降,过去不清楚这究竟是因为总上下文变长了(Length),还是因为文档数量变多了(Distraction/Integration)。

Shahar Levy,Nir Mazor,Lihi Shalmon,Michael Hassid,Gabriel Stanovsky
Hebrew University of Jerusalemraglong-contextanalysisAnthologyDBLP
5
泛读FindingsEMNLP 2025

FroM: Frobenius Norm-Based Data-Free Adaptive Model Merging

这篇论文要解决的是模型合并对数据的依赖问题。现有 adaptive model merging 往往需要验证集或少量校准数据来估计不同模型/参数块的权重,但很多真实场景拿不到可用数据,或者数据本身会泄露隐私和分布偏置。

Zijian Li,Xiaocheng Feng,Huixin Liu,Yichong Huang,Ting Liu,Bing Qin
model-mergingdata-freeAnthologyDBLP
5
泛读FindingsEMNLP 2025

Not All Voices Are Rewarded Equally: Probing and Repairing Reward Models across Human Diversity

这篇论文要解决的是 reward model 对不同人群声音和偏好表达的系统性不公平。标准 reward model 常被当作单一“人类偏好”的代理,但真实人群在措辞、礼貌风格、文化习惯和风险偏好上差异很大,这会让某些群体的合理回答被系统性低估。

Zihao Li,Feihao Fang,Xitong Zhang,Jiaru Zou,Zhining Liu,Wei Xiong,Ziwei Wu,Baoyu Jing,Jingrui He
reward-modelbiasalignmentAnthologyDBLP
5
泛读EMNLP 2025

From Generation to Judgment: Opportunities and Challenges of LLM-as-a-judge

Dawei Li,Bohan Jiang,Liangjie Huang,Alimohammad Beigi,Chengshuai Zhao,Zhen Tan ... 省略 3 位作者 ... ,Tianhao Wu,Kai Shu,Lu Cheng,Huan Liu
llm-as-a-judgeevaluationDOIDBLP
5
泛读EMNLP 2025

DICE: Structured Reasoning in LLMs through SLM-Guided Chain-of-Thought Correction

Yiqi Li,Yusheng Liao,Zhe Chen,Yanfeng Wang,Yu Wang
reasoningcotcorrectionDOIDBLP
5
泛读FindingsEMNLP 2025

Humanity's Last Code Exam: Can Advanced LLMs Conquer Human's Hardest Code Competition?

Xiangyang Li,Xiaopeng Li,Kuicai Dong,Quanhu Zhang,Rongju Ruan,Xinyi Dai,Yasheng Wang,Ruiming Tang
benchmarkcode-generationreasoningAnthologyDBLP
5
泛读FindingsEMNLP 2025

MultiLingPoT: Boosting Mathematical Reasoning in LLMs through Multilingual Program Integration

Nianqi Li,Zujie Liang,Siyu Yuan,Jiaqing Liang,Feng Wei,Yanghua Xiao
reasoningmultilingualprogram-synthesisAnthologyDBLP
5
泛读EMNLP 2025

RAG-Zeval: Enhancing RAG Responses Evaluator through End-to-End Reasoning and Ranking-Based Reinforcement Learning

Kun Li,Yunxiang Li,Tianhua Zhang,Hongyin Luo,Xixin Wu,James R. Glass,Helen M. Meng
ragreinforcement-learningevaluationDOIDBLP
5
泛读FindingsEMNLP 2025

How Does Knowledge Selection Help Retrieval Augmented Generation?

在 RAG 里"知识选择"(对检索回来的段落做重排/筛选)到底什么时候真正有用,这点一直靠经验。大家知道加个 reranker 有时涨点有时不涨,但缺乏系统的机制解释。

Xiangci Li,Jessica Ouyang
University of Texas at Dallasragretrievalknowledge-selectionAnthologyDBLP
5
泛读EMNLP 2025

Co-Evolving LLMs and Embedding Models via Density-Guided Preference Optimization for Text Clustering

文本聚类里 LLM 和 embedding 模型一直是分工割裂的:embedding 给向量、LLM 给语义判断,两者谁都不帮对方。作者想让它们在同一个训练循环里互相提升。

Zetong Li,Qinliang Su,Minhua Huang,Yin Yang
Sun Yat-sen Universityembeddingpreference-optimizationco-trainingDOIDBLP
5
泛读FindingsEMNLP 2025

Mitigating Object Hallucinations in MLLMs via Multi-Frequency Perturbations

MLLM 物体幻觉很多缓解方法都在语义层下手(prompt、decoding、对齐 loss),但视觉 token 本身在不同频段承载的信息差异很大,这个角度没人系统用过。

Shuo Li,Jiajun Sun,Guodong Zheng,Xiaoran Fan,Yujiong Shen,Yi Lu ... 省略 3 位作者 ... ,Tao Ji,Tao Gui,Qi Zhang,Xuanjing Huang
mllmhallucinationrobustnessAnthologyDBLP
5
泛读FindingsEMNLP 2025

Instance-level Randomization: Toward More Stable LLM Evaluations

LLM 评测不稳定是老问题,prompt 模板、选项顺序、few-shot 示例都会让分数抖好几个点。现有做法要么多次重跑平均,要么固定一套模板,但两条路都有局限。

Yiyang Li,Yonghuang Wu,Ying Luo,Liangtai Sun,Zishu Qin,Lin Qiu,Xuezhi Cao,Xunliang Cai
evaluationstabilitybenchmarkAnthologyDBLP
5
泛读FindingsEMNLP 2025

Curse of Knowledge: Your Guidance and Provided Knowledge are biasing LLM Judges in Complex Evaluation

Weiyuan Li,Xintao Wang,Siyu Yuan,Rui Xu,Jiangjie Chen,Qingqing Dong,Yanghua Xiao,Deqing Yang
llm-judgeevaluationbiasAnthologyDBLP
5
泛读EMNLP 2025

RethinkMCTS: Refining Erroneous Thoughts in Monte Carlo Tree Search for Code Generation

Qingyao Li,Wei Xia,Xinyi Dai,Kounianhua Du,Weiwen Liu,Yasheng Wang,Ruiming Tang,Yong Yu,Weinan Zhang
code-generationmctsreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

Topic-Guided Reinforcement Learning with LLMs for Enhancing Multi-Document Summarization

Chuyuan Li,Austin Xu,Shafiq Joty,Giuseppe Carenini
reinforcement-learningsummarizationtopic-controlAnthologyDBLP
5
泛读FindingsEMNLP 2025

EmByte: Decomposition and Compression Learning for Small yet Private NLP

Shenglan Li,Jia Xu,Mengjiao Zhang
compressionprivacysmall-modelAnthologyDBLP
5
泛读FindingsEMNLP 2025

Forget for Get: A Lightweight Two-phase Gradient Method for Knowledge Editing in Large Language Models

Yanhong Li,Min Yang,Xiping Hu,Chengming Li
knowledge-editinggradient-methodmodel-editingAnthologyDBLP
5
泛读EMNLP 2025

RAG-Instruct: Boosting LLMs with Diverse Retrieval-Augmented Instructions

这篇论文要解决的问题是:现有 RAG 指令数据通常覆盖面窄、检索场景单一,导致模型学到的是固定套路而不是稳定的 retrieval-aware 行为。作者想用更丰富的检索增强指令来提升模型在真实检索链路中的泛化。

Wanlong Liu,Junying Chen,Ke Ji,Li Zhou,Wenyu Chen,Benyou Wang
raginstruction-tuningsynthetic-dataDOIDBLP
5
泛读FindingsEMNLP 2025

Thinking Before You Speak: A Proactive Test-time Scaling Approach

这篇论文解决的是测试时扩展(test-time scaling)常常发生在“生成之后再反思”阶段,导致错误思路已经外化,修正成本高且容易越修越偏。作者提出把额外算力前移到回答之前,让模型先形成内部计划或判断,再决定怎么说。

Cong Liu,Wenchang Chai,Hejun Wu,Yan Pan,Pengxu Wei,Liang Lin
test-time-scalingreasoningAnthologyDBLP
5
泛读EMNLP 2025

CogDual: Enhancing Dual Cognition of LLMs via Reinforcement Learning with Implicit Rule-Based Rewards

这篇论文要解决的是 LLM 往往难以同时兼顾快而直觉的响应与慢而规则化的推理,即所谓 dual cognition 的失衡问题。传统 RLHF/RLAIF 通常只优化单一结果分数,难以显式鼓励模型形成这两种认知模式的协同。

Cheng Liu,Yifei Lu,Fanghua Ye,Jian Li,Xingyu Chen,Feiliang Ren,Zhaopeng Tu,Xiaolong Li
rlrewardreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

FlexQuant: A Flexible and Efficient Dynamic Precision Switching Framework for LLM Quantization

Fangxin Liu,Zongwu Wang,JinHong Xia,Junping Zhao,Shouren Zhao,Jinjin Li,Jian Liu,Li Jiang,Haibing Guan
quantizationdynamic-precisioninference-efficiencyAnthologyDBLP
5
泛读FindingsEMNLP 2025

LoRE-Merging: Exploring Low-Rank Estimation For Large Language Model Merging

Zehua Liu,Han Wu,Yuxuan Yao,Xiaojin Fu,Ruifeng She,Xiongwei Han,Tao Zhong,Mingxuan Yuan
model-merginglow-rankllm-compressionAnthologyDBLP
5
泛读EMNLP 2025

GRASP: Replace Redundant Layers with Adaptive Singular Parameters for Efficient Model Compression

Kainan Liu,Yong Zhang,Ning Cheng,Zhitao Li,Shaojun Wang,Jing Xiao
model-compressionlow-ranklayer-pruningDOIDBLP
5
泛读FindingsEMNLP 2025

Data or Language Supervision: What Makes CLIP Better than DINO?

Yiming Liu,Yuhui Zhang,Dhruba Ghosh,Ludwig Schmidt,Serena Yeung-Levy
clipdinodata-supervisionAnthologyDBLP
5
泛读FindingsEMNLP 2025

ViFT: Towards Visual Instruction-Free Fine-tuning for Large Vision-Language Models

Zikang Liu,Kun Zhou,Wayne Xin Zhao,Dawei Gao,Yaliang Li,Ji-Rong Wen
vlminstruction-freefine-tuningAnthologyDBLP
5
泛读FindingsEMNLP 2025

On the Perception Bottleneck of VLMs for Chart Understanding

Junteng Liu,Weihao Zeng,Xiwen Zhang,Yijun Wang,Zifei Shan,Junxian He
vlmperceptionevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

How Jailbreak Defenses Work and Ensemble? A Mechanistic Investigation

Zhuohan Long,Siyuan Wang,Shujun Liu,Yuhang Lai
jailbreaksafetymechanistic-analysisAnthologyDBLP
5
泛读EMNLP 2025

Learn and Unlearn: Addressing Misinformation in Multilingual LLMs

Taiming Lu,Philipp Koehn
unlearningmisinformationmultilingualDOIDBLP
5
泛读FindingsEMNLP 2025

X-Boundary: Establishing Exact Safety Boundary to Shield LLMs from Jailbreak Attacks without Compromising Usability

Xiaoya Lu,Dongrui Liu,Yi Yu,Luxin Xu,Jing Shao
safetyjailbreakalignmentAnthologyDBLP
5
泛读EMNLP 2025

Paired by the Teacher: Turning Unpaired Data into High-Fidelity Pairs for Low-Resource Text Generation

Yen-Ju Lu,Thomas Thebaud,Laureano Moro-Velázquez,Najim Dehak,Jesús Villalba
data-synthesislow-resourcedistillationDOIDBLP
5
泛读EMNLP 2025

Representation Potentials of Foundation Models for Multimodal Alignment: A Survey

Jianglin Lu,Hailing Wang,Yi Xu,Yizhou Wang,Kuo Yang,Yun Fu
surveymultimodal-alignmentrepresentationDOIDBLP
5
泛读EMNLP 2025

ViPE: Visual Perception in Parameter Space for Efficient Video-Language Understanding

Shichen Lu,Tongtian Yue,Longteng Guo,Handong Li,Xingjian He,Si Liu,Jing Liu
video-languageefficient-architecturevisual-perceptionDOIDBLP
5
泛读EMNLP 2025

GlotEval: A Test Suite for Massively Multilingual Evaluation of Large Language Models

Hengyu Luo,Zihao Li,Joseph Attieh,Sawal Devkota,Ona de Gibert,Xu Huang ... 省略 5 位作者 ... ,Mengjie Wang,Samea Yusofi,Fei Yuan,Jörg Tiedemann
evaluationmultilingualbenchmarkDOIDBLP
5
泛读EMNLP 2025

Task-Aware Resolution Optimization for Visual Large Language Models

Weiqing Luo,Zhen Tan,Yifan Li,Xinyu Zhao,Kwonjoon Lee,Behzad Dariush,Tianlong Chen
vlmresolutionoptimizationDOIDBLP
5
泛读FindingsEMNLP 2025

MMATH: A Multilingual Benchmark for Mathematical Reasoning

Wenyang Luo,Wayne Xin Zhao,Jing Sha,Shijin Wang,Ji-Rong Wen
benchmarkmath-reasoningmultilingualAnthologyDBLP
5
泛读EMNLP 2025

Gamma-Guard: Lightweight Residual Adapters for Robust Guardrails in Large Language Models

Lijia Lv,Yuanshu Zhao,Guan Wang,Xuehai Tang,Jie Wen,Jizhong Han,Songlin Hu
safetyguardrailsadapterDOIDBLP
5
泛读EMNLP 2025

Memorization ≠ Understanding: Do Large Language Models Have the Ability of Scenario Cognition?

Boxiang Ma,Ru Li,Yuanlong Wang,Hongye Tan,Xiaoli Li
memorizationunderstandingevaluationDOIDBLP
5
泛读EMNLP 2025

LATTE: Learning to Think with Vision Specialists

通用视觉语言模型(VLM)依赖单一视觉编码器(如 CLIP),在处理细粒度 OCR、密集检测等专业视觉任务时,受限于分辨率和特征提取偏好,能力严重不足。

Zixian Ma,Jianguo Zhang,Zhiwei Liu,Jieyu Zhang,Juntao Tan,Manli Shu ... 省略 2 位作者 ... ,Huan Wang,Caiming Xiong,Ranjay Krishna,Silvio Savarese
Salesforce ResearchvlmreasoningvisionDOIDBLP
5
泛读EMNLP 2025

Pruning the Paradox: How CLIP's Most Informative Heads Enhance Performance While Amplifying Bias

CLIP 模型的注意力头中,对下游任务性能贡献最大的头,往往也是编码社会偏见最严重的头,导致模型性能与公平性难以兼得。

Avinash Madasu,Vasudev Lal,Phillip Howard
Intel Labsclipattention-headspruningDOIDBLP
5
泛读FindingsEMNLP 2025

Embedding-Free RAG

传统 RAG 强依赖独立的稠密向量模型(Embedding Models),增加了系统部署复杂度,且向量表征可能存在语义压缩瓶颈,导致检索不精准。

Jessica Maghakian,Raunak Sinha,Max Schettewi,Gunkirat Kaur
ragretrievalembedding-freeAnthologyDBLP
5
泛读EMNLP 2025

Supervised Attention Mechanism for Low-quality Multimodal Data

多模态模型在面对现实中低质量、缺失或噪声严重的模态数据时,标准交叉注意力机制容易被噪声误导,导致性能发生灾难性下降。

Sijie Mai,Shiqin Han,Haifeng Hu
South China University of Technologymultimodaldata-qualityattentionDOIDBLP
5
泛读EMNLP 2025

Explaining Differences Between Model Pairs in Natural Language through Sample Learning

评估两个 LLM 的差异通常只能看 Benchmark 的聚合分数,无法用自然语言直观解释它们在微观行为、风格偏好或语义理解上的具体区别。

Advaith Malladi,Rakesh R. Menon,Yuvraj Jain,Shashank Srivastava
UNC Chapel Hillmodel-comparisoninterpretabilitysample-learningDOIDBLP
5
泛读FindingsEMNLP 2025

FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models

现有的长文本事实性评估方法(如基于 LLM 裁判)往往是确定性的,难以处理复杂长文本中事实的概率性、模糊性和逻辑依赖,容易产生误判。

Radu Marinescu,Debarun Bhattacharjya,Junkyu Lee,Tigran T. Tchrakian,Javier Carnerero-Cano,Yufang Hou,Elizabeth M. Daly,Alessandra Pascale
IBM Researchfactualityevaluationprobabilistic-modelingAnthologyDBLP
5
泛读FindingsEMNLP 2025

Minimal Ranks, Maximum Confidence: Parameter-efficient Uncertainty Quantification for LoRA

对 LLM 进行不确定性量化(UQ)通常需要全参数的贝叶斯网络或模型集成,计算和显存开销极大,无法应用于千亿参数的大模型。

Patryk Marszalek,Klaudia Balazy,Jacek Tabor,Tomasz Kusmierczyk
Jagiellonian UniversitylorauncertaintypeftAnthologyDBLP
5
泛读EMNLP 2025

LLMs Don't Know Their Own Decision Boundaries: The Unreliability of Self-Generated Counterfactual Explanations

这篇论文的核心判断是:LLM 自己生成的反事实解释并不能可靠刻画它自己的决策边界。过去很多解释性做法默认模型既能给答案也能说明“什么改动会让答案改变”,但这种自我报告可能只是语言上合理,而不是机制上真实。

Harry Mayne,Ryan Othniel Kearns,Yushi Yang,Andrew M. Bean,Eoin D. Delaney,Chris Russell,Adam Mahdi
self-explanationcounterfactualreliabilityDOIDBLP
5
泛读EMNLP 2025

NileChat: Towards Linguistically Diverse and Culturally Aware LLMs for Local Communities

这篇工作要解决的是:主流 LLM 对本地社区语言变体和文化语境覆盖不足,导致即使“会该语言”,也未必能在真实社区场景中给出得体、准确、文化上可接受的回答。过去这类问题常被用更大通用模型或英文化数据绕开,但这通常不能真正补足地域语言与文化知识缺口。

Abdellah El Mekki,Houdaifa Atou,Omer Nacar,Shady Shehata,Muhammad Abdul-Mageed
multilinguallow-resourcecontinual-pretrainDOIDBLP
5
泛读EMNLP 2025

Easy Dataset: A Unified and Extensible Framework for Synthesizing LLM Fine-Tuning Data from Unstructured Documents

这篇工作针对的是一个常见但工程上很痛的问题:如何把非结构化文档稳定地转成高质量、可扩展的 LLM 微调数据。过去很多团队靠人工规则、一次性脚本或让 LLM 直接自由生成训练样本,这常导致格式不统一、覆盖不可控、质量参差且难扩展到新文档类型。

Ziyang Miao,Qiyu Sun,Jingyuan Wang,Yuchen Gong,Yaowei Zheng,Shiqi Li,Richong Zhang
synthetic-datasftdata-pipelineDOIDBLP
5
泛读IndustryEMNLP 2025

Group Preference Alignment: Customizing LLM Responses from In-Situ Conversations Only When Needed

这篇论文要解决的是:如何仅凭真实会话中的 in-situ 信号,为不同用户群体定制 LLM 回答,而且只在确有必要时才进行对齐调整。过去个性化或群体对齐往往依赖显式偏好数据、额外标注或对所有输入一律改写,这既贵,也容易把通用能力过度扭曲。

Ishani Mondal,Jack W. Stokes,Sujay Kumar Jauhar,Longqi Yang,Mengting Wan,Xiaofeng Xu,Xia Song,Jordan Lee Boyd-Graber,Jennifer Neville
preferencealignmentpersonalizationDOIDBLP
5
泛读FindingsEMNLP 2025

RTTC: Reward-Guided Collaborative Test-Time Compute

Juan Pablo Muñoz,Jinjie Yuan
test-time-computerewardAnthologyDBLP
5
泛读FindingsEMNLP 2025

Can Multiple Responses from an LLM Reveal the Sources of Its Uncertainty?

Yang Nan,Pengfei He,Ravi Tandon,Han Xu
uncertaintysamplingAnthologyDBLP
5
泛读FindingsEMNLP 2025

Distributional Surgery for Language Model Activations

Bao Nguyen,Binh T. Nguyen,Duy Nguyen,Viet Anh Nguyen
activationsrepresentationAnthologyDBLP
5
泛读EMNLP 2025

Embedding Domain Knowledge for Large Language Models via Reinforcement Learning from Augmented Generation

Chaojun Nie,Jun Zhou,Guanxiang Wang,Shisong Wu,Zichen Wang
rldomain-adaptationragDOIDBLP
5
泛读FindingsEMNLP 2025

CoTD-PO: Chain-of-Thought Distillation with Preference Optimization

Lujie Niu,Haochen Sun,Fangkun Zhao,Sheng Chen,Zimeng Bai,Jiawei Zhang,Caixia Yuan,Xiaojie Wang
distillationcotdpoAnthologyDBLP
5
泛读FindingsEMNLP 2025

Beyond Fixed-Length Calibration for Post-Training Compression of LLMs

Jaehoon Oh,Dokwan Oh
quantizationpost-training-compressioncalibrationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Beyond Linear Steering: Unified Multi-Attribute Control for Language Models

Narmeen Oozeer,Luke Marks,Fazl Barez,Amir Abdullah
steeringcontrollabilityrepresentation-engineeringAnthologyDBLP
5
泛读EMNLP 2025

Token-level Proximal Policy Optimization for Query Generation

Yichen Ouyang,Lu Wang,Fangkai Yang,Pu Zhao,Chenghua Huang,Jianfeng Liu ... 省略 5 位作者 ... ,Saravan Rajmohan,Weiwei Deng,Dongmei Zhang,Feng Sun
ppotoken-levelrlDOIDBLP
5
泛读FindingsEMNLP 2025

Likelihood Variance as Text Importance for Resampling Texts to Map Language Models

Momose Oyama,Ryo Kishino,Hiroaki Yamagiwa,Hidetoshi Shimodaira
likelihooddata-importanceresamplingAnthologyDBLP
5
泛读EMNLP 2025

Child-Directed Language Does Not Consistently Boost Syntax Learning in Language Models

Francesca Padovani,Jaap Jumelet,Yevgen Matusevych,Arianna Bisazza
syntax-learningchild-directed-speechlanguage-acquisitionDOIDBLP
5
泛读EMNLP 2025

Probability Distribution Collapse: A Critical Bottleneck to Compact Unsupervised Neural Grammar Induction

Jinwook Park,Kangil Kim
grammar-inductionprobability-collapseunsupervisedDOIDBLP
5
泛读FindingsEMNLP 2025

Riemannian Optimization for LoRA on the Stiefel Manifold

JuneYoung Park,Minjae Kang,Seongbae Lee,Haegang Lee,Seongwan Kim,Jaeho Lee
loraoptimizationpeftAnthologyDBLP
5
泛读EMNLP 2025

When Big Models Train Small Ones: Label-Free Model Parity Alignment for Efficient Visual Question Answering using Small VLMs

Abhirama Subramanyam Penamakuri,Navlika Singh,Piyush Arora,Anand Mishra
knowledge-distillationvlmvqaDOIDBLP
5
泛读EMNLP 2025

MUZO: Leveraging Multiple Queries and Momentum for Zeroth-Order Fine-Tuning of Large Language Models

Yuezhang Peng,Yuxin Liu,Fei Wen,Xie Chen
zeroth-order-optimizationfine-tuningefficient-trainingDOIDBLP
5
泛读EMNLP 2025

Firewall Routing: Blocking Leads to Better Hybrid Inference for LLMs

Runyu Peng,Yunhua Zhou,Kai Lv,Yang Gao,Qipeng Guo,Xipeng Qiu
inference-efficiencymodel-routinghybrid-inferenceDOIDBLP
5
泛读EMNLP 2025

SilVar: Speech-Driven Multimodal Model for Reasoning Visual Question Answering and Object Localization

Tan-Hanh Pham,Hoang-Nam Le,Phu-Vinh Nguyen,Chris Ngo,Truong-Son Hy
multimodalspeech-visionvqaDOIDBLP
5
泛读IndustryEMNLP 2025

How to Fine-Tune Safely on a Budget: Model Adaptation Using Minimal Resources

Anh C. Pham,Mihir Thalanki,Michael Sun,Aditya Chaloo,Ankita Gupta,Tian Xia,Aditya Mate,Ehi Nosakhare,Soundararajan Srinivasan
efficient-fine-tuningsafety-alignmentresource-constrainedDOIDBLP
5
泛读FindingsEMNLP 2025

Think Twice, Generate Once: Safeguarding by Progressive Self-Reflection

解决 LLM 在生成过程中容易被越狱或输出不安全内容的问题。传统的静态对齐容易被绕过,而后置过滤(Post-hoc filtering)成本高且延迟大。

Hoang Phan,Victor Li,Qi Lei
self-reflectionsafeguardinginference-time-interventionAnthologyDBLP
5
泛读EMNLP 2025

MR. Judge: Multimodal Reasoner as a Judge

多模态大模型(MLLM)的评估高度依赖人工,缺乏一个可靠的、能像 GPT-4 那样作为裁判的开源多模态评测模型。

Renjie Pi,Haoping Bai,Qibin Chen,Xiaoming Simon Wang,Jiulong Shan,Xiaojiang Liu,Meng Cao
llm-as-a-judgemultimodal-evaluationreasoningDOIDBLP
5
泛读EMNLP 2025

Pointing to a Llama and Call it a Camel: On the Sycophancy of Multimodal Large Language Models

多模态大模型(VLM)在面对用户提供的错误文本提示时,往往会放弃图像中的真实视觉证据,转而迎合用户的错误预设(即阿谀奉承 Sycophancy)。

Renjie Pi,Kehao Miao,Li Peihang,Runtao Liu,Jiahui Gao,Jipeng Zhang,Xiaofang Zhou
vlmsycophancymodel-behaviorDOIDBLP
5
泛读IndustryEMNLP 2025

RAGulator: Lightweight Out-of-Context Detectors for Grounded Text Generation

RAG 系统生成的文本容易出现脱离检索上下文的幻觉(Out-of-Context),而用大模型做实时检测成本过高且延迟大。

Ian Poey,Jiajun Liu,Qishuai Zhong
raghallucinationgroundingDOIDBLP
5
泛读FindingsEMNLP 2025

Dissecting Persona-Driven Reasoning in Language Models via Activation Patching

LLM 在被赋予特定角色(Persona)时,其推理逻辑和输出风格会发生改变,但这种角色扮演是如何在模型内部电路上生效的尚不清楚。

Ansh Poonia,Maeghal Jain
interpretabilityactivation-patchingpersonaAnthologyDBLP
5
泛读FindingsEMNLP 2025

Bhaasha, Bhāṣā, Zaban: A Survey for Low-Resourced Languages in South Asia - Current Stage and Challenges

南亚低资源语言在 LLM 预训练和评测中严重缺乏高质量数据和基准,导致多语言模型在这些语种上表现不佳。

Sampoorna Poria,Xiaolei Huang
low-resourcedata-qualitymultilingualAnthologyDBLP
5
泛读FindingsEMNLP 2025

Can Role Vectors Affect LLM Behaviour?

除了通过 Prompt 让模型进行角色扮演,能否直接在隐空间(Latent Space)通过向量干预来改变模型的行为模式?

Daniele Potertì,Andrea Seveso,Fabio Mercorio
representationpersonabehavior-controlAnthologyDBLP
5
泛读FindingsEMNLP 2025

Dynamic Evaluation for Oversensitivity in LLMs

现有的安全对齐(如 RLHF)往往导致模型过度敏感(Oversensitivity),对正常的、略带边界的 prompt 也会拒绝回答,而静态评测无法捕捉这一动态交互缺陷。

Sophia Xiao Pu,Sitao Cheng,Xin Eric Wang,William Yang Wang
UCSBalignmentevaluationrobustnessAnthologyDBLP
5
泛读FindingsEMNLP 2025

When Models Reason in Your Language: Controlling Thinking Language Comes at the Cost of Accuracy

在多语言推理中,强制模型用非英语(用户的母语)输出思维链(CoT)会导致准确率下降。过去通常假设模型具备跨语言的等效推理能力,但实际上思考语言与推理性能强绑定。

Jirui Qi,Shan Chen,Zidi Xiong,Raquel Fernández,Danielle S. Bitterman,Arianna Bisazza
multilingualreasoningcotAnthologyDBLP
5
泛读EMNLP 2025

SwiftKV: Fast Prefill-Optimized Inference with Knowledge-Preserving Model Transformation

这篇工作要解决的是 LLM 推理里最难被系统优化同时又最影响首 token 延迟的 prefill 阶段。现有方法多半靠内核优化、KV cache 工程或近似注意力提速,但这些方案通常不改模型本身,因此在长上下文和大 batch 下仍受限于算量;而更激进的结构改动又容易破坏模型知识与输出分布。

Aurick Qiao,Zhewei Yao,Samyam Rajbhandari,Yuxiong He
kv-cacheprefillinferenceDOIDBLP
5
泛读EMNLP 2025

IG-Pruning: Input-Guided Block Pruning for Large Language Models

这篇工作要解决的是 LLM 剪枝在实际推理里常见的一个失配:静态剪枝对所有输入一刀切,因此平均有效但对具体样本往往剪错位置,既浪费可裁剪冗余,也放大难例退化。这个问题现在值得重做,是因为部署侧越来越在意按样本自适应的算力分配,而 block-level 结构化剪枝比非结构化更可落地。

Kangyu Qiao,Shaolei Zhang,Yang Feng
pruningcompressioninferenceDOIDBLP
5
泛读EMNLP 2025

Reinforced Query Reasoners for Reasoning-intensive Retrieval Tasks

这篇工作要解决的是 reasoning-intensive retrieval 里“检索查询不会想”的问题。现有 RAG 和 dense retrieval 多数直接把原问题或轻量 reformulation 送进检索器,但在多跳、约束组合或需要中间假设的任务上,这种查询通常信息不完整,导致后续生成再强也吃不到对的证据。

Xubo Qin,Jun Bai,Jiaqi Li,Zixia Jia,Zilong Zheng
retrievalreinforcement-learningreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

TreeBoN: Enhancing Inference-Time Alignment with Speculative Tree-Search and Best-of-N Sampling

这篇工作要解决的是 inference-time alignment 常见的两难:best-of-N 能提高回答质量,但采样代价高;树搜索能更系统地探索,但 rollout 成本和价值评估开销更高。现有方法经常在质量和时延之间只能取其一,尤其在推理型任务上更明显。

Jiahao Qiu,Yifu Lu,Yifan Zeng,Jiacheng Guo,Jiayi Geng,Chenhao Zhu ... 省略 2 位作者 ... ,Huazheng Wang,Kaixuan Huang,Yue Wu,Mengdi Wang
best-of-ntree-searchalignmentAnthologyDBLP
5
泛读EMNLP 2025

Superpose Task-specific Features for Model Merging

这篇工作要解决的是模型合并里任务特征相互覆盖的问题。现有 merge 方法常把多个 finetuned 权重简单线性组合、稀疏选择或按重要性重加权,但不同任务更新往往共享参数方向,结果是某个任务增强的同时把另一个任务冲掉,尤其在任务差异较大时更明显。

Haiquan Qiu,You Wu,Dong Li,Jianmin Guo,Quanming Yao
model-mergingfinetuningrepresentationDOIDBLP
5
泛读EMNLP 2025

On the Same Wavelength? Evaluating Pragmatic Reasoning in Language Models across Broad Concepts

这篇工作要解决的是我们对语言模型 pragmatic reasoning 的评估过窄,导致模型看起来会“理解言外之意”,但其实可能只是在少数经典设定上记住了模式。过去不少研究集中在某个特定语用现象,如 implicature、reference 或 speaker-listener game,而缺少跨广泛概念的一致测量。

Linlu Qiu,Cedegao E. Zhang,Joshua B. Tenenbaum,Yoon Kim,Roger P. Levy
pragmatic-reasoningevaluationlanguage-understandingDOIDBLP
5
泛读FindingsEMNLP 2025

Uplift-RAG: Uplift-Driven Knowledge Preference Alignment for Retrieval-Augmented Generation

这篇工作要解决的是 RAG 对知识的使用常常“检到了但没用对”,而现有对齐方法多偏好输出风格、帮助性或事实性,缺少针对“哪些检索知识应被偏好吸收”的专门优化。结果是模型可能引用无关片段、忽视关键证据,或者在外部知识与参数记忆冲突时做出次优选择。

Changle Qu,Sunhao Dai,Hengyi Cai,Yiyang Cheng,Jun Xu,Shuaiqiang Wang,Dawei Yin
ragalignmentpreference-optimizationAnthologyDBLP
5
泛读EMNLP 2025

CIE: Controlling Language Model Text Generations Using Continuous Signals

Vinay Samuel,Harshita Diddee,Yiming Zhang,Daphne Ippolito
controllable-generationsteeringDOIDBLP
5
泛读EMNLP 2025

What's in a prompt? Language models encode literary style in prompt embeddings

Raphaël Sarfati,Haley Moller,Toni J. B. Liu,Nicolas Boullé,Christopher J. Earls
prompt-embeddingstyleanalysisDOIDBLP
5
泛读EMNLP 2025

Mitigating Hallucinations in Vision-Language Models through Image-Guided Head Suppression

Sreetama Sarkar,Yue Che,Alex Gavin,Peter Anthony Beerel,Souvik Kundu
vlmhallucinationattentionDOIDBLP
5
泛读FindingsEMNLP 2025

Out-of-Context Reasoning in Large Language Models

Jonathan Shaki,Emanuele La Malfa,Michael J. Wooldridge,Sarit Kraus
reasoningout-of-contextllmAnthologyDBLP
5
泛读EMNLP 2025

A Head to Predict and a Head to Question: Pre-trained Uncertainty Quantification Heads for Hallucination Detection in LLM Outputs

Artem Shelmanov,Ekaterina Fadeeva,Akim Tsvigun,Ivan Tsvigun,Zhuohan Xie,Igor Kiselev ... 省略 2 位作者 ... ,Artem Vazhentsev,Mrinmaya Sachan,Preslav Nakov,Timothy Baldwin
uncertainty-quantificationhallucinationpretrained-headsDOIDBLP
5
泛读OutstandingEMNLP 2025

MiCRo: Mixture Modeling and Context-aware Routing for Personalized Preference Learning

个性化偏好学习(personalized preference learning)的难点在于不同用户的偏好分布差异大,单一 reward model 或对齐策略难以同时满足多样化需求。以往方法要么忽略个性化,要么需要为每个用户单独训练,成本高。

Jingyan Shen,Jiarui Yao,Rui Yang,Yifan Sun,Feng Luo,Rui Pan,Tong Zhang,Han Zhao
preference-learningroutingpersonalizationDOIDBLP
5
泛读FindingsEMNLP 2025

LAVa: Layer-wise KV Cache Eviction with Dynamic Budget Allocation

长上下文推理中 KV cache 占用大量显存,需要淘汰部分 KV 条目来降低开销。现有 KV cache eviction 方法通常对所有层使用统一的淘汰预算,忽略了不同层对注意力信息的需求差异。

Yiqun Shen,Song Yuan,Zhengze Zhang,Xiaoliang Wang,Daxin Jiang,Cam-Tu Nguyen
kv-cacheinference-efficiencydynamic-allocationAnthologyDBLP
5
泛读FindingsEMNLP 2025

UniSpeaker: A Unified Approach for Multimodality-driven Speaker Generation

说话人生成(speaker generation)涉及多种模态信息(语音特征、面部外观、说话风格等),现有方法通常针对单一模态或任务设计,缺乏统一框架。本文提出一个多模态驱动的统一说话人生成方法。

Zhengyan Sheng,Zhihao Du,Heng Lu,Shiliang Zhang,Zhen-Hua Ling
USTCaudio-generationmultimodalAnthologyDBLP
5
泛读EMNLP 2025

Do LLMs Behave as Claimed? Investigating How LLMs Follow Their Own Claims using Counterfactual Questions

LLM 在被问到自身行为时会给出声明(claims),但这些声明是否与其实际行为一致?本文通过反事实问题来检验 LLM 的自我声明与实际行为之间的一致性。

Haochen Shi,Shaobo Li,Guoqing Chao,Xiaoliang Shi,Wentao Chen,Zhenzhou Ji
llm-behaviorcounterfactual-reasoningself-consistencyDOIDBLP
5
泛读EMNLP 2025

Reasoning under Uncertainty: Efficient LLM Inference via Unsupervised Confidence Dilution and Convergent Adaptive Sampling

LLM 在推理时通常采用固定的采样策略(如恒定的 Temperature),导致在简单步骤上浪费算力,而在复杂、高不确定性的步骤上又缺乏足够的探索。

Zhenning Shi,Yijia Zhu,Yi Xie,Junhan Shi,Guorui Xie,Haotian Zhang,Yong Jiang,Congcong Miao,Qing Li
inference-samplinguncertaintyadaptive-samplingDOIDBLP
5
泛读FindingsEMNLP 2025

Large Vision-Language Model Alignment and Misalignment: A Survey Through the Lens of Explainability

大型视觉-语言模型(LVLM)在对齐过程中经常出现严重的失调现象(如物体幻觉、跨模态干扰),但现有的对齐研究多停留在经验层面,缺乏从模型内部机制(可解释性)视角的深入剖析。

Dong Shu,Haiyan Zhao,Jingyu Hu,Weiru Liu,Ali Payani,Lu Cheng,Mengnan Du
vlmalignmentsurveyAnthologyDBLP
5
泛读IndustryEMNLP 2025

LLMs on a Budget? Say HOLA

标准的参数高效微调(PEFT)方法如 LoRA 虽然降低了显存占用,但在极端预算受限的场景下仍显笨重,且低秩矩阵分解难以捕捉复杂任务所需的高阶权重交互。

Zohaib Hasan Siddiqui,Jiechao Gao,Ebad Shabbir,Mohammad Anas Azeez,Rafiq Ali,Gautam Siddharth Kashyap,Usman Naseem
parameter-efficient-tuningmodel-compressionDOIDBLP
5
泛读EMNLP 2025

TFDP: Token-Efficient Disparity Audits for Autoregressive LLMs via Single-Token Masked Evaluation

这篇工作要解决的是:如何更省 token 地做 autoregressive LLM 的 disparity audit,也就是检测模型在不同群体、属性或提示条件下的系统性差异。传统审计往往依赖完整生成或大量采样,成本高且方差大;因此用单 token masked evaluation 替代长生成,是一个很实际的评测效率问题。

Inderjeet Singh,Ramya Srinivasan,Roman Vainshtein,Hisashi Kojima
biasevaluationtoken-efficiencyDOIDBLP
5
泛读EMNLP 2025

COCO-Tree: Compositional Hierarchical Concept Trees for Enhanced Reasoning in Vision-Language Models

这篇工作关注 VLM 在组合推理上的薄弱点,尤其是模型能识别对象但难以稳定处理层级概念和多概念组合关系。以往很多方法靠更多指令数据或更强 backbone 硬顶性能,但没有显式组织概念结构;因此作者提出 COCO-Tree 这类概念树,目标是把“知道什么”变成“按层级和组合关系去推理”。

Sanchit Sinha,Guangzhi Xiong,Aidong Zhang
vlmreasoningcompositionalityDOIDBLP
5
泛读EMNLP 2025

Steering LLM Reasoning Through Bias-Only Adaptation

这篇工作要解决的是:能否在不大改模型权重的情况下,稳定地引导 LLM 的推理路径。过去常见做法是 LoRA、全量微调或 prompt engineering,但前两者成本不低,后者控制力度有限且不稳定;bias-only adaptation 试图用最小参数改动实现更可控的 reasoning steering。

Viacheslav Sinii,Alexey Gorbatovski,Artem Cherepanov,Boris Shaposhnikov,Nikita Balagansky,Daniil Gavrilov
reasoningsteeringadapterDOIDBLP
5
泛读FindingsEMNLP 2025

SteerVLM: Robust Model Control through Lightweight Activation Steering for Vision Language Models

这篇工作关注 VLM 的模型控制问题:如何在不重训模型的情况下,对视觉语言模型的输出风格、行为边界或安全属性进行稳健控制。现有控制手段常见于 prompt、解码约束或较重的微调,但在多模态场景下往往不够稳定,且容易被视觉输入扰动破坏。

Anushka Sivakumar,Andrew Zhang,Zaber Ibn Abdul Hakim,Christopher Thomas
vlmsteeringactivationAnthologyDBLP
5
泛读FindingsEMNLP 2025

MUG-Eval: A Proxy Evaluation Framework for Multilingual Generation Capabilities in Any Language

这篇工作要解决的是多语言生成评测的覆盖和成本问题,尤其是很多语言缺少高质量人工基准,导致“支持任何语言”的生成能力很难可靠衡量。过去通常只测少数高资源语言,或者依赖昂贵人工评审;MUG-Eval 试图提供一个可扩展的 proxy evaluation framework。

Seyoung Song,Seogyeong Jeong,Eunsu Kim,Jiho Jin,Dongkwan Kim,Jay Shin,Alice Oh
multilingualevaluationgenerationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Smart-Searcher: Incentivizing the Dynamic Knowledge Acquisition of LLMs via Reinforcement Learning

Huatong Song,Jinhao Jiang,Wenqing Tian,Zhipeng Chen,Yuhuan Wu,Jiahao Zhao,Yingqian Min,Xin Zhao,Lei Fang,Ji-Rong Wen
reinforcement-learningtool-usesearchAnthologyDBLP
5
泛读FindingsEMNLP 2025

Safety Through Reasoning: An Empirical Study of Reasoning Guardrail Models

Makesh Narsimhan Sreedhar,Traian Rebedea,Christopher Parisien
safetyreasoningguardrailAnthologyDBLP
5
泛读EMNLP 2025

Unlearning vs. Obfuscation: Are We Truly Removing Knowledge?

Guangzhi Sun,Potsawee Manakul,Xiao Zhan,Mark J. F. Gales
unlearningsafetyDOIDBLP
5
泛读EMNLP 2025

An Empirical Study of LLM Reasoning Ability Under Strict Output Length Constraint

Yi Sun,Han Wang,Jiaqiang Li,Jiacheng Liu,Xiangyu Li,Hao Wen ... 省略 1 位作者 ... ,Huiwen Zheng,Yan Liang,Yuanchun Li,Yunxin Liu
reasoninglengthempiricalDOIDBLP
5
泛读FindingsEMNLP 2025

Hard Negatives, Hard Lessons: Revisiting Training Data Quality for Robust Information Retrieval with LLMs

Nandan Thakur,Crystina Zhang,Xueguang Ma,Jimmy Lin
retrievaldata-qualityhard-negativesAnthologyDBLP
5
泛读FindingsEMNLP 2025

Harry Potter is Still Here! Probing Knowledge Leakage in Targeted Unlearned Large Language Models

现有的 LLM 知识遗忘(Unlearning)方法是否真正从权重中擦除了知识,还是仅仅掩盖了提取路径。

Bang Trinh Tran To,Thai Le
unlearningknowledge-leakagememorizationAnthologyDBLP
5
泛读EMNLP 2025

Investigating Neurons and Heads in Transformer-based LLMs for Typographical Errors

Transformer 模型对输入中的拼写错误(Typos)具有惊人的鲁棒性,但其将破碎的 Token 映射回正确语义的内部机制尚不明确。

Kohei Tsuji,Tatsuya Hiraoka,Yuchang Cheng,Eiji Aramaki,Tomoya Iwakura
interpretabilityneuronsattention-headsDOIDBLP
5
泛读EMNLP 2025

RRInf: Efficient Influence Function Estimation via Ridge Regression for Large Language Models and Text-to-Image Diffusion Models

计算影响函数(Influence Functions)以进行预训练数据归因时,海量参数导致的逆海森矩阵计算成本过高,难以在 LLM 和 Diffusion 模型上扩展。

Zhuozhuo Tu,Cheng Chen,Yuxuan Du
influence-functiondata-attributiondiffusionDOIDBLP
5
泛读EMNLP 2025

ViLBench: A Suite for Vision-Language Process Reward Modeling

视觉语言模型(VLM)缺乏细粒度、多步骤推理的评估基准,导致难以训练和评估用于多模态复杂任务的过程奖励模型(PRM)。

Haoqin Tu,Weitao Feng,Hardy Chen,Hui Liu,Xianfeng Tang,Cihang Xie
process-reward-modelvlmbenchmarkDOIDBLP
5
泛读EMNLP 2025

Quantifying Logical Consistency in Transformers via Query-Key Alignment

评估 LLM 生成内容的逻辑一致性通常依赖脆弱的黑盒 Prompting,缺乏一种能反映模型内部状态的白盒量化指标。

Eduard Tulchinskii,Laida Kushnareva,Anastasia Voznyuk,Andrei Andriiainen,Irina Piontkovskaya,Evgeny Burnaev,Serguei Barannikov
logical-consistencyquery-key-alignmentinterpretabilityDOIDBLP
5
泛读EMNLP 2025

Unconditional Truthfulness: Learning Unconditional Uncertainty of Large Language Models

LLM 往往表现出过度自信,其输出的不确定性(Uncertainty)极易受到 Prompt 措辞的干扰,导致在未知领域产生严重幻觉。

Artem Vazhentsev,Ekaterina Fadeeva,Rui Xing,Gleb Kuzmin,Ivan Lazichny,Alexander Panchenko,Preslav Nakov,Timothy Baldwin,Maxim Panov,Artem Shelmanov
uncertaintytruthfulnesscalibrationDOIDBLP
5
泛读EMNLP 2025

ProLongVid: A Simple but Strong Baseline for Long-context Video Instruction Tuning

这篇论文要解决的是:长上下文视频指令微调缺少足够强、足够简单的基线,导致很多方法改动很多却难判断真实增益来自哪里。视频 instruction tuning 常见问题是序列极长、帧冗余高、监督信号稀疏,复杂方法未必比扎实的工程基线更值得投入。

Rui Wang,Bohao Li,Xiyang Dai,Jianwei Yang,Yi-Ling Chen,Zhen Xing ... 省略 1 位作者 ... ,Dongdong Chen,Xipeng Qiu,Zuxuan Wu,Yu-Gang Jiang
video-llmlong-contextinstruction-tuningDOIDBLP
5
泛读FindingsEMNLP 2025

SecDecoding: Steerable Decoding for Safer LLM Generation

这篇论文要解决的是:现有安全对齐主要靠训练时约束,但推理时用户意图和风险级别是动态的,模型需要更细粒度、可调的安全生成控制。传统做法要么把安全都塞进 SFT/RLHF,要么用硬规则拦截;前者缺可控性,后者往往过于粗糙。

Jiayou Wang,Rundong Liu,Yue Hu,Huijia Wu,Zhaofeng He
safe-decodingalignmentsteeringAnthologyDBLP
5
泛读EMNLP 2025

LVLMs are Bad at Overhearing Human Referential Communication

这篇论文要解决的是:LVLM 在视觉问答里看起来会“理解场景”,但在更接近人类自然交流的 referential communication 场景中,尤其是旁听式理解(overhearing),可能并不具备同等能力。以往评测多是明确指令、明确目标的单轮任务,回避了真实交流里隐含说话人意图、共同背景和指称消解的难点。

Zhengxiang Wang,Weiling Li,Panagiotis Kaliosis,Owen Rambow,Susan Brennan
lvlmreferential-communicationevaluationDOIDBLP
5
泛读EMNLP 2025

Not All Parameters Are Created Equal: Smart Isolation Boosts Fine-Tuning Performance

这篇论文要解决的是:参数高效微调通常默认所有可训练参数的重要性相近,但实际上不同参数对迁移和稳定性的贡献差异很大。过去很多方法只是减少可训练参数数量,却没有认真处理“该训练哪些、该隔离哪些”这个更细的问题。

Yao Wang,Di Liang,Minlong Peng
fine-tuningparameter-efficientadaptationDOIDBLP
5
泛读EMNLP 2025

From Parameters to Performance: A Data-Driven Study on LLM Structure and Development

从大量已发布 LLM 的结构参数出发,用数据驱动的方式回答'在给定预算下模型应该长什么样'这一实证问题。过去这个问题要么靠零散的 scaling law,要么靠厂商经验,缺乏把宽度/深度/头数/FFN 比例等超参联合考量的系统归纳。

Suqing Wang,Zuchao Li,Luohe Shi,Bo Du,Hai Zhao,Yun Li,Qianren Wang
武汉大学上海交通大学scalingmodel-analysisarchitectureDOIDBLP
5
泛读FindingsEMNLP 2025

Diving into Mitigating Hallucinations from a Vision Perspective for Large Vision-Language Models

LVLM 幻觉的根源究竟在'视觉侧'还是'语言侧'——这篇明确把矛头指向视觉表示本身不够强,而不是 LLM 解码端 overconfident。过去缓解方案大量集中在对比解码、RLHF、指令数据上,对视觉 tokenizer/编码器这一侧的干预相对少。

Weihang Wang,Xinhao Li,Ziyue Wang,Yan Pang,Jielei Zhang,Peiyi Li,Qiang Zhang,Longwen Gao
hallucinationvlmvisionAnthologyDBLP
5
泛读FindingsEMNLP 2025

FIER: Fine-Grained and Efficient KV Cache Retrieval for Long-context LLM Inference

长上下文推理下 KV cache 又大又慢,现有稀疏检索方法要么粒度太粗(按 block 选)导致召回差,要么太细(token 级 attention)开销扛不住。

Dongwei Wang,Zijie Liu,Song Wang,Yuxin Ren,Jianing Deng,Jingtong Hu,Tianlong Chen,Huanrui Yang
kv-cachelong-contextinferenceAnthologyDBLP
5
泛读EMNLP 2025

Advancing Fine-Grained Visual Understanding with Multi-Scale Alignment in Multi-Modal Models

MLLM 对细粒度视觉(小物体、局部属性、文字)理解偏弱,核心是单一尺度的视觉特征不足以同时覆盖全局语义和局部细节。

Wei Wang,Zhaowei Li,Qi Xu,Linfeng Li,Yiqing Cai,Botian Jiang,Hang Song,Xingcan Hu,Pengyu Wang,Li Xiao
visual-understandingalignmentmultiscaleDOIDBLP
5
泛读EMNLP 2025

FacLens: Transferable Probe for Foreseeing Non-Factuality in Fact-Seeking Question Answering of Large Language Models

能否在模型生成答案之前就预判它会不会胡编——即 pre-generation 层面的非事实性检测。现有幻觉检测大多在生成之后做(采样一致性、自检),代价高且晚。

Yanling Wang,Haoyang Li,Hao Zou,Jing Zhang,Xinlei He,Qi Li,Ke Xu
factualityhallucinationprobeDOIDBLP
5
泛读FindingsEMNLP 2025

VideoLLM Knows When to Speak: Enhancing Time-Sensitive Video Comprehension with Video-Text Duet Interaction Format

VideoLLM 在时间敏感任务(什么时候该说话、定位事件、在线解说)上表现差,根源是训练格式——把整段视频一次性喂给模型再输出——本身就没教模型'时机感'。

Yueqian Wang,Xiaojun Meng,Yuxuan Wang,Jianxin Liang,Jiansheng Wei,Huishuai Zhang,Dongyan Zhao
华为诺亚方舟实验室video-llmtime-sensitivevideoAnthologyDBLP
5
泛读EMNLP 2025

HD-PiSSA: High-Rank Distributed Orthogonal Adaptation

PiSSA 类(基于主奇异分量初始化的 LoRA)在单卡上已经比普通 LoRA 好,但秩仍受单卡限制。作者要在分布式场景下把有效秩拉高,同时避免多卡适配器之间的冗余。

Yiding Wang,Fanxu Meng,Xuefeng Zhang,Fan Jiang,Pingzhi Tang,Muhan Zhang
lorafine-tuningadaptationDOIDBLP
5
泛读EMNLP 2025

LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization

Xujia Wang,Yunjia Qi,Bin Xu
fine-tuningsubnethigh-rankDOIDBLP
5
泛读FindingsEMNLP 2025

Improving LLM Reasoning through Interpretable Role-Playing Steering

Anyi Wang,Dong Shu,Yifan Wang,Yunpu Ma,Mengnan Du
reasoningsteeringrole-playingAnthologyDBLP
5
泛读FindingsEMNLP 2025

Flexible Thinking for Multimodal Emotional Support Conversation via Reinforcement Learning

Fanfan Wang,Xiangqing Shen,Jianfei Yu,Rui Xia
reinforcement-learningmultimodalconversationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Unveiling Multimodal Processing: Exploring Activation Patterns in Multimodal LLMs for Interpretability and Efficiency

多模态大模型在处理图文混合输入时,其内部的激活模式(Activation Patterns)和模态路由机制尚不透明,导致计算存在冗余。

Chuan Wu,Meng Su,Youxuan Fang,Shaolin Zhu
multimodal-llminterpretabilityactivation-analysisAnthologyDBLP
5
泛读FindingsEMNLP 2025

Mitigating Visual Knowledge Forgetting in MLLM Instruction-tuning via Modality-decoupled Gradient Descent

这篇论文要解决的是:MLLM 在指令微调后常出现视觉知识遗忘,语言对齐变强了,但视觉理解被冲淡。这个问题以前多用混合图文数据或保守学习率来缓解,但这类办法往往只是在吞掉遗忘,而没有直接处理图像梯度和语言梯度相互干扰的根因。

Junda Wu,Yuxin Xiong,Xintong Li,Yu Xia,Ruoyu Wang,Yu Wang ... 省略 2 位作者 ... ,Ryan A. Rossi,Lina Yao,Jingbo Shang,Julian J. McAuley
vlminstruction-tuningcatastrophic-forgettingAnthologyDBLP
5
泛读EMNLP 2025

Thought calibration: Efficient and confident test-time scaling

这篇论文要解决的是:test-time scaling 想提高答案质量,通常要生成更多推理样本,但代价高且置信度不可靠。过去做法要么盲目多采样,要么靠启发式停止规则,缺少一个能同时控制计算开销和答案可信度的校准机制。

Menghua Wu,Cai Zhou,Stephen Bates,Tommi S. Jaakkola
test-time-scalingchain-of-thoughtcalibrationDOIDBLP
5
泛读FindingsEMNLP 2025

BcQLM: Efficient Vision-Language Understanding with Distilled Q-Gated Cross-Modal Fusion

这篇论文要解决的是:视觉语言理解模型如何在保持跨模态交互能力的同时降低计算成本。传统 cross-attention 融合往往效果好但开销大,尤其在高分辨率视觉 token 或长文本条件下,融合层容易成为延迟和显存瓶颈。

Sike Xiang,Shuang Chen,Amir Atapour-Abarghouei
vlmcross-modal-fusiondistillationAnthologyDBLP
5
泛读FindingsEMNLP 2025

LongWeave: A Long-Form Generation Benchmark Bridging Real-World Relevance and Verifiability

长文本生成评测长期卡在'真实任务'和'可验证性'二选一的矛盾里。已有 benchmark 要么是开放式写作(靠 LLM-as-judge,打分不稳),要么是人造的合成任务(可自动判分但脱离真实需求),作者想做一个既贴近真实场景又能自动可验证的长文生成 benchmark。

Zikai Xiao,Fei Huang,Jianhong Tu,Jianhui Wei,Wen Ma,Yuxuan Zhou,Jian Wu,Bowen Yu,Zuozhu Liu,Junyang Lin
long-form-generationbenchmarkverifiabilityAnthologyDBLP
5
泛读EMNLP 2025

Whisper-UT: A Unified Translation Framework for Speech and Text

Whisper 原本只做 ASR+X→En 翻译,要做多方向语音翻译和文本翻译就得分别接不同模型或重训。作者想在 Whisper 框架里统一语音翻译、文本翻译、跨方向翻译四种任务。

Cihan Xiao,Matthew Wiesner,Debashish Chakraborty,Reno Kriz,Keith Cunningham,Kenton Murray,Kevin Duh,Luis Tavarez-Arce,Paul McNamee,Sanjeev Khudanpur
speech-translationunified-frameworkmultimodalDOIDBLP
5
泛读EMNLP 2025

From A and B to A+B: Can Large Language Models Solve Compositional Math Problems?

LLM 在能解决子问题 A 和 B 的前提下,能不能自动解决 A+B 的组合问题?现有数学 benchmark 大多测单一技能,无法隔离'组合泛化'这个能力。

Xisheng Xiao,Hanlin Zhao
math-reasoningcompositionalityevaluationDOIDBLP
5
泛读FindingsEMNLP 2025

Reveal and Release: Iterative LLM Unlearning with Self-generated Data

LLM unlearning(让模型忘记指定知识)目前要么靠访问原训练数据构造反例,要么做一次性梯度操作,效果不稳。作者想在不依赖原始遗忘数据的情况下做稳定的 unlearning。

Linxi Xie,Xin Teng,Shichang Ke,Hongyi Wen,Shenji Wan
unlearningself-generated-datasafetyAnthologyDBLP
5
泛读FindingsEMNLP 2025

MAFMO: Multi-modal Adaptive Fusion with Meta-template Optimization for Vision-Language Models

VLM 在小样本/下游任务适配时,提示模板(prompt template)和多模态特征融合方式通常是手工或固定的,换任务就掉点。作者想让模板和融合都自适应。

Mingrui Xie,Lulu Xu,Junliang Du
vlmadaptive-fusionmultimodalAnthologyDBLP
5
泛读IndustryEMNLP 2025

Think-Search-Patch: A Retrieval-Augmented Reasoning Framework for Repository-Level Code Repair

Bojian Xiong,Yikun Lei,Xikai Liu,Shaowei Zhang,Pengyun Zhu,Yan Liu ... 省略 4 位作者 ... ,Yan Gao,Yi Wu,Yao Hu,Deyi Xiong
code-repairrepository-levelragDOIDBLP
5
泛读EMNLP 2025

UNComp: Can Matrix Entropy Uncover Sparsity? - A Compressor Design from an Uncertainty-Aware Perspective

Jing Xiong,Jianghan Shen,Fanghua Ye,Chaofan Tao,Zhongwei Wan,Jianqiao Lu ... 省略 2 位作者 ... ,Zhijiang Guo,Min Yang,Lingpeng Kong,Ngai Wong
model-compressionsparsityefficiencyDOIDBLP
5
泛读EMNLP 2025

OBLIVIATE: Robust and Practical Machine Unlearning for Large Language Models

Xiaoyu Xu,Minxin Du,Qingqing Ye,Haibo Hu
unlearningsafetypost-trainingDOIDBLP
5
泛读FindingsEMNLP 2025

Dense Retrievers Can Fail on Simple Queries: Revealing The Granularity Dilemma of Embeddings

Liyan Xu,Zhenlin Su,Mo Yu,Jiangnan Li,Fandong Meng,Jie Zhou
retrievalembeddingsevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

Self-Ensemble: Mitigating Confidence Distortion for Large Language Models

Zicheng Xu,Guanchu Wang,Guangyao Zheng,Yu-Neng Chuang,Alex Szalay,Xia Hu,Vladimir Braverman
calibrationself-ensembleuncertaintyAnthologyDBLP
5
泛读EMNLP 2025

Alignment for Efficient Tool Calling of Large Language Models

Hongshen Xu,Zihan Wang,Zichen Zhu,Lei Pan,Xingyu Chen,Shuai Fan,Lu Chen,Kai Yu
tool-usealignmentefficiencyDOIDBLP
5
泛读EMNLP 2025

Linguistic Neuron Overlap Patterns to Facilitate Cross-lingual Transfer on Low-resource Languages

Yuemei Xu,Kexin Xu,Jian Zhou,Ling Hu,Lin Gui
cross-lingualneuronstransferDOIDBLP
5
泛读EMNLP 2025

MMLU-ProX: A Multilingual Benchmark for Advanced Large Language Model Evaluation

Weihao Xuan,Rui Yang,Heli Qi,Qingcheng Zeng,Yunze Xiao,Aosong Feng ... 省略 20 位作者 ... ,Lei Ma,Edison Marrese-Taylor,Shijian Lu,Yusuke Iwasawa
multilingualbenchmarkevaluationDOIDBLP
5
泛读EMNLP 2025

Seeing is Believing, but How Much? A Comprehensive Analysis of Verbalized Calibration in Vision-Language Models

Weihao Xuan,Qingcheng Zeng,Heli Qi,Junjue Wang,Naoto Yokoya
vlmcalibrationmultimodal-understandingDOIDBLP
5
泛读EMNLP 2025

TruthTorchLM: A Comprehensive Library for Predicting Truthfulness in LLM Outputs

Duygu Nur Yaldiz,Yavuz Faruk Bakman,Sungmin Kang,Alperen Özis,Hayrettin Eren Yildiz,Mitash Ashish Shah ... 省略 2 位作者 ... ,Alfy Samuel,Daben Liu,Sai Praneeth Karimireddy,Salman Avestimehr
truthfulnesshallucinationevaluationDOIDBLP
5
泛读EMNLP 2025

Hidden in Plain Sight: Reasoning in Underspecified and Misspecified Scenarios for Multimodal LLMs

Qianqi Yan,Hongquan Li,Shan Jiang,Yang Zhao,Xinze Guan,Ching-Chen Kuo,Xin Eric Wang
multimodal-reasoningrobustnessevaluationDOIDBLP
5
泛读EMNLP 2025

LMR-BENCH: Evaluating LLM Agent's Ability on Reproducing Language Modeling Research

Shuo Yan,Ruochen Li,Ziming Luo,Zimu Wang,Daoyang Li,Liqiang Jing ... 省略 5 位作者 ... ,Ziyang Zhang,Mian Zhang,Zhiyu Chen,Xinya Du
agentbenchmarklanguage-modelingDOIDBLP
5
泛读EMNLP 2025

From Long to Lean: Performance-aware and Adaptive Chain-of-Thought Compression via Multi-round Refinement

Jianzhi Yan,Le Liu,Youcheng Pan,Shiwei Chen,Zike Yuan,Yang Xiang,Buzhou Tang
cotcompressionreasoningDOIDBLP
5
泛读EMNLP 2025

Addressing Tokenization Inconsistency in Steganography and Watermarking Based on Large Language Models

大模型的水印和隐写技术通常依赖于在 token 级别修改 logits,但由于 BPE 等子词分词器的“分词不一致性”(前缀改变导致后续文本的分词边界发生变化),会导致解码错误或水印失效。这个问题在理论研究中常被忽略,但在实际部署中是致命的。

Ruiyi Yan,Yugo Murawaki
Kyoto UniversitytokenizerwatermarkingsteganographyDOIDBLP
5
泛读FindingsEMNLP 2025

Confusion is the Final Barrier: Rethinking Jailbreak Evaluation and Investigating the Real Misuse Threat of LLMs

当前的越狱(Jailbreak)评估指标过度惩罚了模型,将模型输出的“困惑”或无意义的顺从(如格式上答应但内容胡言乱语)也算作越狱成功,导致真实的安全威胁被严重高估。

Yu Yan,Sheng Sun,Zhe Wang,Yijun Lin,Zenghao Duan,Zhifei Zheng,Min Liu,Zhiyi Yin,Jianping Zhang
jailbreaksafetyevaluationAnthologyDBLP
5
泛读EMNLP 2025

MUSE: MCTS-Driven Red Teaming Framework for Enhanced Multi-Turn Dialogue Safety in Large Language Models

多轮对话的安全红队测试(Red Teaming)极其困难,因为随着对话轮数增加,攻击空间呈指数级膨胀,传统的单轮启发式或 RL 攻击很难在深层对话树中找到突破口。

Siyu Yan,Long Zeng,Xuecheng Wu,Chengcheng Han,Kongcheng Zhang,Chong Peng,Xuezhi Cao,Xunliang Cai,Chenjuan Guo
red-teamingmctssafetyDOIDBLP
5
泛读EMNLP 2025

Code to Think, Think to Code: A Survey on Code-Enhanced Reasoning and Reasoning-Driven Code Intelligence in LLMs

代码数据能提升大模型的通用逻辑推理能力,而通用推理技术(如 CoT)也能反哺代码生成。这种双向促进现象在业界广为人知,但缺乏系统的理论梳理和实证归纳。

Dayu Yang,Tianyang Liu,Daoan Zhang,Antoine Simoulin,Xiaoyi Liu,Yuwei Cao ... 省略 1 位作者 ... ,Xin Qian,Grey Yang,Jiebo Luo,Julian J. McAuley
surveycodereasoningDOIDBLP
5
泛读EMNLP 2025

UNCLE: Benchmarking Uncertainty Expressions in Long-Form Generation

这篇工作要解决的是:长文本生成中的不确定性表达缺少系统 benchmark,导致模型是‘不知道但说得像知道’还是‘知道自己不确定’很难被量化比较。现有评测多关注事实正确性或校准分数,但对长段落中不确定性的语言形式、位置分布和任务适配性覆盖不足。

Ruihan Yang,Caiqi Zhang,Zhisong Zhang,Xinting Huang,Dong Yu,Nigel Collier,Deqing Yang
uncertaintylong-form-generationbenchmarkDOIDBLP
5
泛读EMNLP 2025

LaMDAgent: An Autonomous Framework for Post-Training Pipeline Optimization via LLM Agents

这篇工作要解决的是:LLM 的 post-training pipeline 很复杂,包含数据配方、超参、评测、迭代决策等大量人工试错,导致优化效率低且难复现。过去 AutoML 更多关注模型或超参搜索,但对现代 LLM 后训练这种多阶段、闭环、带评估反馈的流程支持不足。

Taro Yano,Yoichi Ishibashi,Masafumi Oyamada
post-trainingpipelineagentDOIDBLP
5
泛读EMNLP 2025

CaKE: Circuit-aware Editing Enables Generalizable Knowledge Learners

这篇工作要解决的是:知识编辑往往能改对单个事实,但泛化差、容易副作用大,难以让模型学成可复用的‘知识学习器’。过去很多 editing 方法更像局部修补:命中目标快,但一离开模板或关系邻域就不稳。

Yunzhi Yao,Jizhan Fang,Jia-Chen Gu,Ningyu Zhang,Shumin Deng,Huajun Chen,Nanyun Peng
knowledge-editingcircuitsDOIDBLP
5
泛读FindingsEMNLP 2025

SLIM: Subtrajectory-Level Elimination for More Effective Reasoning

这篇工作要解决的是:长链推理训练里,整条错误轨迹常被整体丢弃或整体降权,导致很多局部正确但后续跑偏的中间步骤没有被利用。这样做的问题是监督粒度太粗,模型学不到‘前半段是对的、错在后半段哪里’。

Xifeng Yao,Chengyuan Ma,Dongyu Lang,Yinhao Ni,Zhiwei Xu,Huarui Xie ... 省略 1 位作者 ... ,Guang Shen,Dandan Tu,Yi Bai,Changzheng Zhang
reasoningrltrajectoryAnthologyDBLP
5
泛读FindingsEMNLP 2025

TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use

这篇工作要解决的是:LLM 的 tool use 训练往往按任务名或数据集名组织,忽略了任务背后的能力特征,导致训练覆盖不均、迁移性差。过去常见做法是堆更多工具调用样本,但没有明确回答‘哪些任务特征才是决定工具使用泛化的关键变量’。

Junjie Ye,Yilong Wu,Sixian Li,Yuming Yang,Zhiheng Xi,Tao Gui ... 省略 2 位作者 ... ,Peng Wang,Zhongchao Shi,Jianping Fan,Zhengyin Du
tool-usetraining-frameworkAnthologyDBLP
5
泛读FindingsEMNLP 2025

CPO: Addressing Reward Ambiguity in Role-playing Dialogue via Comparative Policy Optimization

这篇工作要解决的是:角色扮演对话中的奖励存在歧义,同一回复可能在角色一致性、帮助性、安全性和风格沉浸感之间相互拉扯,导致单一标量奖励难以稳定优化。现有方法若直接做 RLHF/DPO,往往把这些维度压成一个模糊偏好,训练信号容易互相冲突。

Jing Ye,Rui Wang,Yuchuan Wu,Victor Ma,Feiteng Fang,Fei Huang,Yongbin Li
role-playingrlrewardAnthologyDBLP
5
泛读FindingsEMNLP 2025

SPPD: Self-training with Process Preference Learning Using Dynamic Value Margin

Hao Yi,Qingyang Li,Yulan Hu,Fuzheng Zhang,Di Zhang,Yong Liu
self-trainingpreference-learningprocess-rewardAnthologyDBLP
5
泛读EMNLP 2025

R-TOFU: Unlearning in Large Reasoning Models

Sangyeon Yoon,Wonje Jeung,Albert No
unlearningreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting in Multimodal LLMs

Zeping Yu,Sophia Ananiadou
continual-learningparameter-fusioncatastrophic-forgettingAnthologyDBLP
5
泛读FindingsEMNLP 2025

EvolKV: Evolutionary KV Cache Compression for LLM Inference

Bohan Yu,Yekun Chai
kv-cachecompressioninference-efficiencyAnthologyDBLP
5
泛读EMNLP 2025

PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality

Byeongho Yu,Changhun Lee,Jungyu Jin,Eunhyeok Park
decodingfactualitycontrastive-decodingDOIDBLP
5
泛读FindingsEMNLP 2025

Explainable Chain-of-Thought Reasoning: An Empirical Analysis on State-Aware Reasoning Dynamics

Sheldon Yu,Yuxin Xiong,Junda Wu,Xintong Li,Tong Yu,Xiang Chen,Ritwik Sinha,Jingbo Shang,Julian J. McAuley
chain-of-thoughtinterpretabilityreasoning-dynamicsAnthologyDBLP
5
泛读EMNLP 2025

iTool: Reinforced Fine-Tuning with Dynamic Deficiency Calibration for Advanced Tool Use

Yirong Zeng,Xiao Ding,Yuxian Wang,Weiwen Liu,Yutai Hou,Wu Ning ... 省略 1 位作者 ... ,Duyu Tang,Dandan Tu,Bing Qin,Ting Liu
rlhftool-usefine-tuningDOIDBLP
5
泛读FindingsEMNLP 2025

NUMINA: A Natural Understanding Benchmark for Multi-dimensional Intelligence and Numerical Reasoning Abilities

Changyu Zeng,Yifan Wang,Zimu Wang,Wei Wang,Zhengni Yang,Muyi Bao,Jimin Xiao,Anh Nguyen,Yutao Yue
benchmarkreasoningmathAnthologyDBLP
5
泛读EMNLP 2025

Active Layer-Contrastive Decoding Reduces Hallucination in Large Language Model Generation

Hongxiang Zhang,Hao Chen,Muhao Chen,Tianyi Zhang
decodinghallucinationcontrastive-decodingDOIDBLP
5
泛读EMNLP 2025

An Orthogonal High-Rank Adaptation for Large Language Models

Xin Zhang,Guang-Ze Chen,Shuzhen Li,Zhulin Liu,C. L. Philip Chen,Tong Zhang
peftloraadaptationDOIDBLP
5
泛读EMNLP 2025

Sparse Neurons Carry Strong Signals of Question Ambiguity in LLMs

Zhuoxuan Zhang,Jinhao Duan,Edward Kim,Kaidi Xu
interpretabilitysparse-neuronsambiguityDOIDBLP
5
泛读FindingsEMNLP 2025

SPFT-SQL: Enhancing Large Language Model for Text-to-SQL Parsing by Self-Play Fine-Tuning

Yuhao Zhang,Shaoming Duan,Jinhang Su,Chuanyi Liu,Peiyi Han
self-playtext-to-sqlfine-tuningAnthologyDBLP
5
泛读EMNLP 2025

DrDiff: Dynamic Routing Diffusion with Hierarchical Attention for Breaking the Efficiency-Quality Trade-off

扩散模型在迭代去噪过程中对所有时间步和空间区域分配同等算力,导致推理延迟极高且存在大量冗余计算。

Jusheng Zhang,Yijia Fan,Kaitong Cai,Zimeng Huang,Xiaofei Sun,Jian Wang,Chengpei Tang,Keze Wang
diffusiondynamic-routingattentionDOIDBLP
5
泛读FindingsEMNLP 2025

Bold Claims or Self-Doubt? Factuality Hallucination Type Detection via Belief State

现有的幻觉检测方法过度依赖外部知识库或模型输出的概率置信度,而经过 RLHF 对齐的模型往往存在置信度校准不良(过度自信)的问题。

Dongyu Zhang,Qingqing Hong,Bingxuan Hou,Jiayi Lin,Chenyang Zhang,Jialin Li,Junli Wang
hallucinationbelief-stateinterpretabilityAnthologyDBLP
5
泛读FindingsEMNLP 2025

Sensitivity-LoRA : Low-Load Sensitivity-Based Fine-Tuning for Large Language Models

标准 LoRA 在所有层和模块上分配相同的秩(Rank),忽略了不同层对下游任务的敏感度差异,导致参数利用率次优。

Hao Zhang,Bo Huang,Zhenjia Li,Xi Xiao,Hui Yi Leong,Zumeng Zhang,Xinwei Long,Tianyang Wang,Hao Xu
peftlorafine-tuningAnthologyDBLP
5
泛读EMNLP 2025

ActionStudio: A Lightweight Framework for Data and Training of Large Action Models

大型动作模型(LAM/Agent)的训练缺乏统一的基础设施,不同环境(Web、OS、API)的轨迹数据格式各异,导致数据收集和模型训练的工程壁垒极高。

Jianguo Zhang,Thai Hoang,Ming Zhu,Zuxin Liu,Shiyu Wang,Tulika Awalgaonkar ... 省略 6 位作者 ... ,Shelby Heinecke,Huan Wang,Silvio Savarese,Caiming Xiong
action-modeltraining-dataframeworkDOIDBLP
5
泛读FindingsEMNLP 2025

Staged Knowledge Distillation Through Least-to-Most Prompting: Optimizing Teacher Guidance via Difficulty-Aware Training

在复杂推理任务中,直接将大模型(Teacher)的完整思维链(CoT)蒸馏给小模型(Student)往往效果不佳,因为小模型难以一次性吸收过于复杂的推理逻辑。

Mengxiang Zhang,Lingyuan Liu
knowledge-distillationleast-to-mostcurriculumAnthologyDBLP
5
泛读FindingsEMNLP 2025

Hierarchical Reward Modeling for Fault Localization in Large Code Repositories

这篇工作要解决的是:如何让奖励模型在大规模代码仓库的 fault localization 场景里学到“分层定位”能力,而不是只对单个文件或片段打分。已有方法通常把问题扁平化成候选片段排序,能在小范围代码修复里工作,但遇到跨文件依赖、仓库级调用链和长上下文时,奖励信号会变得稀疏且不稳定,因此值得单独建模。

Jiwei Zhang,Jianxun Lian,Haiming Qin,Mingyang Zhou,Kezhong Lu,Rui Mao,Hao Liao
reward-modelfault-localizationcodeAnthologyDBLP
5
泛读FindingsEMNLP 2025

TIU-Bench: A Benchmark for Evaluating Large Multimodal Models on Text-rich Image Understanding

这篇工作要解决的是:现有多模态评测对 text-rich image understanding 覆盖不足,导致我们很难知道 LMM 在复杂图文混合场景里到底是视觉理解问题、OCR 问题,还是跨模态推理问题。过去很多 benchmark 把图片当自然图像处理,文本密集图像只占很小比例,因此现在单独做 TIU-Bench 是有意义的。

Kun Zhang,Liqiang Niu,Zhen Cao,Fandong Meng,Jie Zhou
benchmarkmllmtext-rich-imagesAnthologyDBLP
5
泛读EMNLP 2025

All Roads Lead to Rome: Graph-Based Confidence Estimation for Large Language Model Reasoning

针对 LLM 推理置信度估计的问题:单条 CoT 的 token 概率或 self-consistency 投票都没有利用多条推理路径之间的结构关系。作者想把多次采样得到的推理路径当成一张图来建模置信度,而不是简单数票。

Caiqi Zhang,Chang Shu,Ehsan Shareghi,Nigel Collier
University of CambridgeMonash UniversityreasoningconfidencegraphsDOIDBLP
5
泛读FindingsEMNLP 2025

DecoupledESC: Enhancing Emotional Support Generation via Strategy-Response Decoupled Preference Optimization

情感支持对话里,策略(要安慰/要建议/要共情)和具体回复内容高度耦合,DPO/SFT 把它们混在一起训练时,策略选择被回复表面形式带偏。作者想把策略层和回复层的偏好学习解耦。

Chao Zhang,Xin Shi,Xueqiao Zhang,Yifan Zhu,Yi Yang,Yawei Luo
preference-optimizationalignmentgenerationAnthologyDBLP
5
泛读EMNLP 2025

Utility-Focused LLM Annotation for Retrieval and Retrieval-Augmented Generation

用 LLM 给检索/RAG 任务自动标注相关性,目前多数做法沿用传统 IR 的 topical relevance 标准,但 RAG 真正关心的是这条文档能否帮下游生成回答正确——两者不一定一致。作者要把标注目标改成 utility(对生成是否有用)而不是 relevance。

Hengran Zhang,Minghao Tang,Keping Bi,Jiafeng Guo,Shihao Liu,Daiting Shi,Dawei Yin,Xueqi Cheng
Chinese Academy of Sciences, ICTannotationretrievalragDOIDBLP
5
泛读EMNLP 2025

REARANK: Reasoning Re-ranking Agent via Reinforcement Learning

Reranker 目前要么是传统 cross-encoder(无推理),要么是 LLM pointwise/listwise 打分(弱推理),都没把重排当成一个需要多步推理的问题。作者用 RL 把 reranker 训成一个会思考的 agent。

Le Zhang,Bo Wang,Xipeng Qiu,Siva Reddy,Aishwarya Agrawal
MilaMcGill UniversityFudan UniversityrlrerankingreasoningDOIDBLP
5
泛读FindingsEMNLP 2025

Recognizing Limits: Investigating Infeasibility in Large Language Models

LLM 面对不可完成(infeasible)的任务时倾向硬编——编造答案或给出看似合理的空洞回答,而不是承认任务做不了。作者要系统研究 LLM 识别不可行任务的能力。

Wenbo Zhang,Zihang Xu,Hengrui Cai
infeasibilityrefusalcalibrationAnthologyDBLP
5
泛读FindingsEMNLP 2025

From Remembering to Metacognition: Do Existing Benchmarks Accurately Evaluate LLMs?

现有 LLM benchmark 大量集中在 '记忆+应用' 层,但 Bloom's taxonomy 的高阶层(analysis/evaluation/metacognition)很少被显式考到。作者问:现在的 benchmark 到底在测什么能力?

Geng Zhang,Yizhou Ying,Sihang Jiang,Jiaqing Liang,Guanglei Yue,Yifei Fu,Hailin Hu,Yanghua Xiao
benchmarkmetacognitionevaluationAnthologyDBLP
5
泛读FindingsEMNLP 2025

ExeSQL: Self-Taught Text-to-SQL Models with Execution-Driven Bootstrapping for SQL Dialects

Jipeng Zhang,Haolin Yang,Kehao Miao,Ruiyuan Zhang,Renjie Pi,Jiahui Gao,Xiaofang Zhou
text-to-sqlself-taughtbootstrappingAnthologyDBLP
5
泛读EMNLP 2025

NOVA-63: Native Omni-lingual Versatile Assessments of 63 Disciplines

Jinyang Zhang,Kexin Yang,Yu Wan,Muyang Ye,Baosong Yang,Fei Huang,Junyang Lin,Dayiheng Liu
benchmarkmultilingualevaluationDOIDBLP
5
泛读EMNLP 2025

MaZO: Masked Zeroth-Order Optimization for Multi-Task Fine-Tuning of Large Language Models

Zhen Zhang,Yifan Yang,Kai Zhen,Nathan Susanj,Athanasios Mouchtaris,Siegfried Kunzmann,Zheng Zhang
zeroth-orderfine-tuningmulti-taskDOIDBLP
5
泛读EMNLP 2025

Continuous-Time Attention: PDE-Guided Mechanisms for Long-Sequence Transformers

Yukun Zhang,Xueqing Zhou
attentionlong-contextarchitectureDOIDBLP
5
泛读FindingsEMNLP 2025

Exploring the Generalizability of Factual Hallucination Mitigation via Enhancing Precise Knowledge Utilization

Siyuan Zhang,Yichi Zhang,Yinpeng Dong,Hang Su
hallucinationknowledge-utilizationAnthologyDBLP
5
泛读EMNLP 2025

Linear-Time Demonstration Selection for In-Context Learning via Gradient Estimation

Ziniu Zhang,Zhenshuo Zhang,Dongyue Li,Lu Wang,Jennifer G. Dy,Hongyang R. Zhang
icldemonstration-selectiongradientDOIDBLP
5
泛读EMNLP 2025

EvolveSearch: An Iterative Self-Evolving Search Agent

LLM 驱动的搜索 agent 在多轮检索中缺乏自我改进能力——它们通常按固定策略检索,无法根据已获取的信息动态调整搜索策略。问题在于如何让 agent 在迭代搜索过程中自主演化查询和推理。

Dingchu Zhang,Yida Zhao,Jialong Wu,Liwen Zhang,Baixuan Li,Wenbiao Yin ... 省略 1 位作者 ... ,Yu-Feng Li,Kewei Tu,Pengjun Xie,Fei Huang
search-agentself-evolvingrlDOIDBLP
5
泛读EMNLP 2025

Sparse Activation Editing for Reliable Instruction Following in Narratives

LLM 在叙事场景中遵循指令的可靠性不足——当指令嵌入在长叙事文本中时,模型容易忽略或错误执行。本文通过稀疏激活编辑来提升叙事场景下的指令遵循能力。

Runcong Zhao,Chengyu Cao,Qinglin Zhu,Xiucheng Lyu,Shun Shao,Lin Gui,Ruifeng Xu,Yulan He
activation-editinginstruction-followingDOIDBLP
5
泛读EMNLP 2025

AdaSteer: Your Aligned LLM is Inherently an Adaptive Jailbreak Defender

对齐后的 LLM 面对 jailbreak 攻击时防御能力不稳定。AdaSteer 的核心观察是:对齐过的 LLM 内部已经包含了区分安全/不安全请求的表示,可以利用这些内在表示来实现自适应防御,而不需要额外的安全分类器或重新训练。

Weixiang Zhao,Jiahe Guo,Yulin Hu,Yang Deng,An Zhang,Xingyu Sui ... 省略 1 位作者 ... ,Yanyan Zhao,Bing Qin,Tat-Seng Chua,Ting Liu
jailbreakalignmentsteeringDOIDBLP
5
泛读FindingsEMNLP 2025

Do We Know What LLMs Don't Know? A Study of Consistency in Knowledge Probing

我们真的知道 LLM 不知道什么吗?现有的知识探测(knowledge probing)方法在不同探测方式下给出不一致的结论——同一个事实,换一种问法模型就可能从'知道'变成'不知道'。本文系统研究了知识探测的一致性问题。

Raoyuan Zhao,Abdullatif Köksal,Ali Modarressi,Michael A. Hedderich,Hinrich Schütze
LMU Munichknowledge-probingconsistencyAnthologyDBLP
5
泛读FindingsEMNLP 2025

Multifaceted Evaluation of Audio-Visual Capability for MLLMs: Effectiveness, Efficiency, Generalizability and Robustness

当前多模态大模型(MLLMs)的评测多局限于单一视觉或单一音频,缺乏对音视频联合建模(Audio-Visual)能力在效率、泛化和鲁棒性等多维度的系统性基准测试。

Yusheng Zhao,Xiao Luo,Junyu Luo,Weizhi Zhang,Zhiping Xiao,Wei Ju,Philip S. Yu,Ming Zhang
audio-visualmllmbenchmarkAnthologyDBLP
5
泛读EMNLP 2025

HELENE: Hessian Layer-wise Clipping and Gradient Annealing for Accelerating Fine-tuning LLM with Zeroth-order Optimization

零阶优化(Zeroth-order Optimization, 如 MeZO)虽能通过避免反向传播大幅降低 LLM 微调显存,但梯度估计方差大、收敛极慢,难以在深层网络中稳定应用。

Huaqin Zhao,Jiaxi Li,Yi Pan,Shizhe Liang,Xiaofeng Yang,Fei Dou,Tianming Liu,Jin Lu
optimizerfine-tuningzeroth-orderDOIDBLP
5
泛读EMNLP 2025

Looking Beyond Text: Reducing Language Bias in Large Vision-Language Models via Multimodal Dual-Attention and Soft-Image Guidance

视觉语言模型(VLMs)在生成时容易过度依赖语言先验(Language Bias),导致“幻觉”——即根据文本惯性捏造图像中不存在的物体,忽视了真实的视觉输入。

Haozhe Zhao,Shuzheng Si,Liang Chen,Yichi Zhang,Maosong Sun,Baobao Chang,Minjia Zhang
Tsinghua Universityvlmlanguage-biasattentionDOIDBLP
5
泛读FindingsEMNLP 2025

Chain of Strategy Optimization Makes Large Language Models Better Emotional Supporter

LLM 在提供情感支持时,往往倾向于给出冰冷的逻辑建议或套路化的安慰,缺乏像人类心理咨询师那样基于策略的深度共情与多轮引导能力。

Weixiang Zhao,Xingyu Sui,Xinyang Han,Yang Deng,Yulin Hu,Jiahe Guo ... 省略 2 位作者 ... ,Shijin Wang,Yanyan Zhao,Bing Qin,Ting Liu
alignmentemotional-supportcotAnthologyDBLP
5
泛读FindingsEMNLP 2025

Dream to Chat: Model-based Reinforcement Learning on Dialogues with User Belief Modeling

传统的对话强化学习(如无模型的 RLHF)难以对用户的长期信念状态(User Belief)和多轮交互动态进行有效建模,导致对话策略短视。

Yue Zhao,Xiaoyu Wang,Dan Wang,Zhonglin Jiang,Qingqing Gu,Teng Chen,Ningyuan Xi,Jinxian Qu,Yong Chen,Luo Ji
rldialogueworld-modelAnthologyDBLP
5
泛读EMNLP 2025

Tiny Budgets, Big Gains: Parameter Placement Strategy in Parameter Super-Efficient Fine-Tuning

在极低参数预算的参数高效微调(PEFT)中,均匀分配可训练参数(如统一的 LoRA rank)无法适应不同 Transformer 层在特征提取上的差异性,导致性能次优。

Jinman Zhao,Xueyan Zhang,Jiaru Li,Jingcheng Niu,Yulan Hu,Erxue Min,Gerald Penn
peftparameter-efficientDOIDBLP
5
泛读FindingsEMNLP 2025

Mitigating Hallucination in Large Vision-Language Models through Aligning Attention Distribution to Information Flow

视觉语言模型(VLM)的幻觉往往源于注意力机制的失效——模型在生成描述时,其注意力权重并没有真正聚焦在对应的视觉信息流上,而是被语言先验主导。

Jianfei Zhao,Feng Zhang,Xin Sun,Chong Feng
vlmhallucinationattentionAnthologyDBLP
5
泛读EMNLP 2025

T2: An Adaptive Test-Time Scaling Strategy for Contextual Question Answering

在上下文问答中,对所有问题使用固定的测试时计算量(Test-Time Compute,如固定的采样次数或检索深度)会导致简单问题算力浪费,而复杂问题算力不足。

Zhengyi Zhao,Shubo Zhang,Zezhong Wang,Huimin Wang,Yutian Zhao,Bin Liang,Yefeng Zheng,Binyang Li,Kam-Fai Wong,Xian Wu
test-time-scalingqaDOIDBLP
5
泛读EMNLP 2025

VoiceCraft-X: Unifying Multilingual, Voice-Cloning Speech Synthesis and Speech Editing

这篇论文要解决的是:如何用一个统一模型同时覆盖多语言语音合成、声音克隆和语音编辑,而不是为每个任务和语言分别训练系统。过去这些能力往往拆成多个模块或多个模型,工程上可行,但会牺牲表示共享、跨任务迁移和部署复杂度。

Zhisheng Zheng,Puyuan Peng,Anuj Diwan,Cong Phuoc Huynh,Xiaohang Sun,Zhu Liu,Vimal Bhat,David Harwath
speech-synthesisvoice-cloningmultilingualDOIDBLP
5
泛读EMNLP 2025

RecBase: Generative Foundation Model Pretraining for Zero-Shot Recommendation

这篇论文要解决的是:推荐系统能否像基础模型一样先做生成式预训练,再在零样本条件下完成推荐,而不是严重依赖任务特定监督和召回/排序流水线。传统推荐方法在各数据集上能做得很强,但迁移弱、冷启动差、任务边界重,这使‘foundation model for recommendation’成为值得单独验证的方向。

Sashuai Zhou,Weinan Gan,Qijiong Liu,Ke Lei,Jieming Zhu,Hai Huang,Yan Xia,Ruiming Tang,Zhenhua Dong,Zhou Zhao
generative-pretrainingrecommendationzero-shotDOIDBLP
5
泛读FindingsEMNLP 2025

LM2Protein: A Structure-to-Token Protein Large Language Model

这篇论文要解决的是:蛋白质大模型通常从序列到结构或功能做预测,但如果直接把结构编码成 token,再做 structure-to-token 的语言建模,是否能更好地捕捉蛋白折叠与设计中的离散规律。过去结构信息常作为连续几何特征喂给图网络或扩散模型,而不是被离散化后纳入 LM 框架。

Chang Zhou,Yuheng Shan,Pengan Chen,Xiangyu Shi,Zikang Wang,Yanting Li,Jiyue Jiang
protein-lmtokenizerstructure-to-tokenAnthologyDBLP
5
泛读EMNLP 2025

BSFA: Leveraging the Subspace Dichotomy to Accelerate Neural Network Training

这篇论文要解决的是:神经网络训练是否存在可利用的子空间二分结构,从而在不显著损伤效果的前提下加速优化。过去训练加速通常靠优化器、低秩近似、梯度压缩或并行系统,而不是假设参数/梯度动态天然落在两类不同作用的子空间里。

WenJie Zhou,Bohan Wang,Wei Chen,Xueqi Cheng
training-accelerationsubspaceoptimizationDOIDBLP
5
泛读FindingsEMNLP 2025

Enhancing LLM Knowledge Learning through Generalization

Mingkang Zhu,Xi Chen,Zhongdao Wang,Bei Yu,Hengshuang Zhao,Jiaya Jia
knowledge-learninggeneralizationcontinual-learningAnthologyDBLP