【Block总结】门控注意力机制,最新注意力机制|即插即用|最佳论文奖论文标题:Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free 作者:Zihan Qiu, Zekun Wang, Bo Zheng 等 会议:NeurIPS 2025(Oral Presentation,入选前1.5%) 论文链接:https://arxiv.org/abs/2505.06708 代码仓库:https://github.com/qiuzh20/gated_at