Skip to content
U-Tokyo Matsuo Lab
U-Tokyo Matsuo Lab
Make the world differentiable
  • Missions & Activities
    • Missions & Activities
    • Members
  • RESEARCH
    • Deep Learning
    • Web & Business
    • Missions & Activities
  • EDUCATION
    • Overview
    • Deep Learning
    • Data Science (GCI)
    • PSI Motivational project
    • 世界モデル・シミュレータ寄付講座
    • AI経営寄付講座
    • Past lecturers and TAs (2020-)
  • PUBLICATIONS
    • Journals
    • Articles
    • International Conferences
    • Japanese Conferences
    • Books
    • Awards
    • Press
    • Invited Lectures
  • STARTUPS
    • STARTUPS
    • 起業クエスト
      • 起業クエスト参加者限定ページ
  • JOIN US
    • Students
    • Recruit
  • CONTACT
  • Language: English
    • 日本語 日本語
    • English English
  • Missions & Activities
    • Missions & Activities
    • Members
  • RESEARCH
    • Deep Learning
    • Web & Business
    • Missions & Activities
  • EDUCATION
    • Overview
    • Deep Learning
    • Data Science (GCI)
    • PSI Motivational project
    • 世界モデル・シミュレータ寄付講座
    • AI経営寄付講座
    • Past lecturers and TAs (2020-)
  • PUBLICATIONS
    • Journals
    • Articles
    • International Conferences
    • Japanese Conferences
    • Books
    • Awards
    • Press
    • Invited Lectures
  • STARTUPS
    • STARTUPS
    • 起業クエスト
      • 起業クエスト参加者限定ページ
  • JOIN US
    • Students
    • Recruit
  • CONTACT
  • Language: English
    • 日本語 日本語
    • English English

Release of Weblab-10B: A 10 Billion-Parameter Bilingual Language Model Supporting Japanese and English

Press Release:

Matsuo Laboratory, The University of Tokyo

Release of Weblab-10B: A 10 Billion-Parameter Bilingual Language Model

Supporting Japanese and English

 

※ The following is a partial English translation of the press release issued on August 22, 2023. Please refer to the Japanese version for the original.

Original: Press Release_2023/08/22

 

Tokyo, [2023/08/22] — Matsuo Laboratory, operating under the Department of Technology Management for Innovation at The University of Tokyo Graduate School of Engineering, announces the release of its large-scale language model, Weblab-10B. Led by Professor Yutaka Matsuo, the laboratory has successfully developed a Large Language Model (LLM) that consists of 10 billion parameters and offers bilingual support for both Japanese and English.

 

Objectives and Future Applications:

 

The lab is primarily focused on advancing artificial intelligence (AI) research and aims to facilitate the technology for industrial applications. The newly developed Weblab-10B model aims to accelerate not just text-based AI but also multimodal applications like image processing and behavior control algorithms for software and robotic platforms. Concurrently, the lab intends to utilize the expertise acquired from this project for educational activities, including course development at the university level.

 

Technical Details:

 

Addressing the imbalance in textual data availability between English and other languages like Japanese, Matsuo Lab has diversified its training data. Weblab-10B has been trained on both English and Japanese datasets, namely The Pile and Japanese-mC4, for the pre-training phase. The post-training or fine-tuning phase involved five distinct datasets: Alpaca (English), Alpaca (Japanese Translation), Flan 2021 (English), Flan CoT (English), and Flan Dialog (English).

 

Benchmarking and Performance Metrics:

 

Notably, despite a lower proportion of Japanese data in the fine-tuning stage, the model showed significant improvements in the JGLUE benchmark evaluation for Japanese, rising from 66% to 78%. These results affirm the model’s efficacy in knowledge transfer between languages. Weblab-10B’s performance stands as a domestic milestone, competitive with other internationally available open-source models.

 

For additional information and model comparison metrics, please refer to the appended open model comparison table.

 

The pre-trained and post-trained models of Weblab-10B developed by us will be released as open model and may not be used for commercial purposes. (See Hugging Face page below)

 

・Pre-trained model

https://huggingface.co/matsuo-lab/weblab-10b

・Post-trained (fine-tuning) model

https://huggingface.co/matsuo-lab/weblab-10b-instruction-sft

 

Notes:

Weblab-10B holds the distinction of being the most advanced open-source Japanese language model to date.

The fine-tuning phase had a relatively smaller portion of Japanese data, underscoring the importance of the model’s capability for cross-lingual knowledge transfer.

 

松尾研 採用情報 ー 研究員・職員のポジションに興味がある方はコチラ ー

松尾研 教職員インタビュー

Author: Kubota

Post navigation

PreviousPrevious post:Mr. Hiroki Furuta has been selected for “FORBES JAPAN 30 UNDER 30 2023”.NextNext post:Matsuo Lab held the LLM Special Boot Camp.

Related posts

Our paper was accepted for Information Processing Society of Japan
September 13, 2023
当研究室の論文が人工知能学会論文誌に採録されました。
September 8, 2023
当研究室の論文が2023年度情報処理学会論文誌ジャーナル特選論文に選出されました。
June 21, 2023
当研究室の研究員が共著となった論文がAdvanced Roboticsに採録されました。
June 2, 2023
当研究室の論文がAdvanced Roboticsに採録されました。
June 2, 2023
当研究室の論文がICML 2023に2件採録されました。
April 26, 2023
MENU
  • Missions & Activities
    • Missions & Activities
    • Members
  • RESEARCH
    • Deep Learning
    • Web & Business
    • Missions & Activities
  • EDUCATION
    • Overview
    • Deep Learning
    • Data Science (GCI)
    • PSI Motivational project
    • 世界モデル・シミュレータ寄付講座
    • AI経営寄付講座
    • Past lecturers and TAs (2020-)
  • PUBLICATIONS
    • Journals
    • Articles
    • International Conferences
    • Japanese Conferences
    • Books
    • Awards
    • Press
    • Invited Lectures
  • STARTUPS
    • STARTUPS
    • 起業クエスト
      • 起業クエスト参加者限定ページ
  • JOIN US
    • Students
    • Recruit
  • CONTACT
  • Language: English
    • 日本語 日本語
    • English English
Copyright ©Matsuo Lab. All Rights Reserved.