• Home
  • ニュース
  • 当研究室の論文6本がEACL 2026に採択されました。
  • 当研究室の論文6本がEACL 2026に採択されました。

    【Main Conference】

    ■書誌情報
    Jerry Huang, Peng Lu, QIUHAO Zeng, Yusuke Iwasawa, Yutaka Matsuo, Sarath Chandar, Edison Marrese-Taylor, Irene Li.
    “Investigating the Multilingual Calibration Effects of Language Model Instruction Tuning”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    Ensuring that deep learning models are well-calibrated in terms of their predictive uncertainty is essential in maintaining their trustworthiness and reliability, yet despite increasing advances in foundation model research, the relationship between such large language models (LLMs) and their calibration remains an open area of research. In this work, we look at a critical gap in the calibration of LLMs within multilingual settings, in an attempt to better understand how the data scarcity can potentially lead to different calibration effects and how commonly used techniques can apply in these settings. Our analysis on two multilingual benchmarks, over 29 and 42 languages respectively, reveals that even in low-resource languages, model confidence can increase significantly after instruction-tuning on high-resource language SFT datasets. However, improvements in accuracy are marginal or non-existent, resulting in mis-calibration, highlighting a critical shortcoming of standard SFT for multilingual languages. Furthermore, we observe that the use of label smoothing to be a reasonable method alleviate this concern, again without any need for low-resource SFT data, maintaining better calibration across all languages. Overall, this highlights the importance of multilingual considerations for both training and tuning LLMs in order to improve their reliability and fairness in downstream use.

    ■書誌情報
    Yunze Xiao, Tingyu He, Lionel Z. WANG, Yiming Ma, Xingyu Song, Xiaohang Xu, Mona T. Diab, Irene Li, Ka Chung Ng.
    “JiraiBench: A Bilingual Benchmark for Evaluating Large Language Models’ Detection of Human risky health behavior Content in Jirai Community”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    In this paper, we present the first cross-lingual dataset that captures a transnational cultural phenomenon, focusing on the Chinese and Japanese “Jirai” subculture and its association with risky health behaviors. Our dataset of more than 15,000 annotated social media posts forms the core of JiraiBench, a benchmark designed to evaluate LLMs on culturally specific content. This unique resource allowed us to uncover an unexpected cross-cultural transfer in which Japanese prompts better handle Chinese content, indicating that cultural context can be more influential than linguistic similarity. Further evidence suggests potential cross-lingual knowledge transfer in fine-tuned models. This work proves the indispensable role of developing culturally informed, cross-lingual datasets for creating effective content moderation tools that can protect vulnerable communities across linguistic borders.

    ■書誌情報
    Shota Takashiro, Takeshi Kojima, Shohei Taniguchi, Yusuke Iwasawa, Yutaka Matsuo.
    “∞-MoE: Generalizing Mixture of Experts to Infinite Experts”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    The Mixture of Experts (MoE) selects a few feed-forward networks (FFNs) per token, achieving an effective trade-off between computational cost and performance. In conventional MoE, each expert is treated as entirely independent, and experts are combined in a discrete space. As a result, when the number of experts increases, it becomes difficult to train each expert effectively. To stabilize training while increasing the number of experts, we propose ∞-MoE that selects a portion of the parameters of large FFNs based on continuous values sampled for each token. By considering experts in a continuous space, this approach allows for an infinite number of experts while maintaining computational efficiency. Experiments show that a GPT-2 Small-based ∞-MoE model, with 129M active and 186M total parameters, achieves comparable performance to a dense GPT-2 Medium with 350M parameters. Adjusting the number of sampled experts at inference time allows for a flexible trade-off between accuracy and speed, with an improvement of up to 2.5% in accuracy over conventional MoE.

    ■書誌情報
    Qi Cao, Andrew Gambardella, Takeshi Kojima, Yutaka Matsuo, Yusuke Iwasawa.
    “Semantic Token Clustering for Efficient Uncertainty Quantification in Large Language Models”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    Large Language Models (LLMs) have demonstrated remarkable capabilities across diverse tasks. However, their limited truthfulness and tendency toward overconfidence constrain their reliability in factual tasks. Uncertainty quantification (UQ) offers a promising approach to identifying potentially unreliable outputs from LLMs. Yet most existing UQ methods rely on repeated sampling or auxiliary models, which substantially increase computational overhead. To address these limitations, we propose an efficient UQ method that leverages semantic information inherently encoded in LLMs. Specifically, we cluster tokens into semantically consistent groups based on embedding similarity and prefix matching, and compute a cluster-based uncertainty score at each decoding step. Our approach requires only a single deterministic generation and does not depend on any auxiliary models. Experiments on multiple datasets and models demonstrate that our method achieves performance comparable to existing baselines while substantially reducing computational overhead.

    【Findings】

    ■書誌情報
    Masaki Sashida, Takeshi Kojima, Yusuke Iwasawa, Yutaka Matsuo.
    “Revealing Redundant Syntax in Large Language Models through Multi-Hop Dependency Paths”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    Prior work on attention–syntax alignment has largely focused on single-hop Universal Dependency edges (DPs). In this paper, we treat short multi-hop dependency paths (MDPs) (e.g., “obl+case”) as first-class units and analyze them alongside DPs. Across three pretrained autoregressive LMs (GPT-2 XL, Llama 3 8B, Qwen3-8B) and one encoder baseline (BERT-large), we extract 2–3 hop MDPs from UD-parsed English and quantify head–relation alignment with an Unlabeled Attachment Score (UAS)–style metric modified for causal masking in decoder-only models. Rank visualizations reveal both overlap and specialization: we observe heads that align with both DPs and MDPs, as well as heads that appear specialized for one route. To test functional relevance, we first identify heads by UAS and then apply an undifferentiated (uniform) attention ablation to those heads; we evaluate the impact on BLiMP and LAMBADA. Ablating the top 10% of all heads shows that MDP-selected heads induce larger drops than DP-selected heads and that the union (“Mix”) of DP- and MDP-selected heads yields the largest drops. For GPT-2 XL, the observed drops are (BLiMP: DP = Δ1.35 pp, MDP = Δ4.81 pp, Mix = Δ7.11 pp; LAMBADA: DP = Δ4.70 pp, MDP = Δ25.17 pp, Mix = Δ32.99 pp), all exceeding size-matched random controls. These results indicate that models can route information consistent with syntactic dependencies via both DP and MDP pathways, with MDPs playing a distinct and measurable role under our interventions.

    ■書誌情報
    Atsushi Shimizu, Shohei Taniguchi, Yutaka Matsuo.
    “Position Encoding with Random Float Sampling Enhances Length Generalization of Transformers”.
    Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2026).
    ■概要
    Length generalization is the ability of language models to maintain performance on inputs longer than those seen during pretraining. In this work, we introduce a simple yet powerful position encoding (PE) strategy, Random Float Sampling (RFS), that generalizes well to lengths unseen during pretraining or fine-tuning. In particular, instead of selecting position indices from a predefined discrete set, RFS uses randomly sampled continuous values, thereby avoiding out-of-distribution (OOD) issues on unseen lengths by exposing the model to diverse indices during training. Since assigning indices to tokens is a common and fundamental procedure in widely used PEs, the advantage of RFS can easily be incorporated into, for instance, the absolute sinusoidal encoding, RoPE, and ALiBi. Experiments corroborate its effectiveness by showing that RFS results in superior performance in length generalization tasks as well as zero-shot commonsense reasoning benchmarks.