LLM-jp LLM-jp

Resources

Presentation materials and surveys used during the study sessions are available to the public. Note that the contents are mainly written in Japanese.

Additionally, we are updating the Overview of Japanese LLMs.

2025-4-22

  • LLM-jp Status Report (Kurohashi) [PDF]
  • Harnessing AI Agents for Real-World Applications ( Kwasi Ankomah / SambaNova, Principle AI Engineer)[PDF]
  • The role of LLM in juris-informatics and a survey of legal control over LLM (Sato) [PDF]
  • mid-training for LLM [PDF]
  • Dialogue WG (Higashinaka) [PDF]
  • Corpus Construction Working Group (Kawahara) [PDF]
  • Real-world Interaction Working Group (Ogata) [PDF]
  • Evaluation and Tuning Working Group (Miyao) [PDF]
  • Safety Working Group (Sekine) [PDF]
  • Multi-modal Working Group (Okazaki) [PDF]
  • Model Building Working Group (Suzuki) [PDF]

2025-3-25

  • LLM-jp Status Report (Kurohashi) Oral

<Evaluation and Turning/Principal Elucidation WG>

  • Are Checklists Really Useful for Automatic Evaluation of Generative Tasks? (Furuhashi) [PDF]
  • Introduction of Open Japanese LLM leaderboard and statistical analysis on evaluation results. (Namgi Han)[PDF]
  • Analyzing the Pretraining of Japanese Large Language Models. (Nishida) [PDF]
  • llm-jp-judge: Japanese LLM-as-a-Judge Evaluation Tool. (Kodama) [PDF]
  • Understanding the Role of Persona and Internal Mechanisms in Large Language Models. (Ozaki)[PDF]
  • How LLMs Learn: Tracing Internal Representations with Sparse Autoencoders. (Inaba)[PDF]
  • A Massive Fine-tuned LLMs from Diverse Models, Tasks, Methods (Harada) [PDF]
  • Comparative analysis of the Geospatial Representations in Large Language Models across Models and Languages (Otake) [PDF]
  • Large-Scale Human Evaluation of LLMs for Japanese(Inoue) [PDF]
  • A Study on Fine-tuning Methods for Balancing Usefulness and Safety in Japanese Large Language Models. (Katsumata)[PDF]

<Multi-modal WG>

  • Developing Japanese CLIP Models Leveraging an Open-weight LLM for Large-scale Dataset Translation. (Sugiura) [PDF]
  • lm-jp-eval-mm: An Evaluation Framework for Evaluating Japanese-centric Vision and Language Model. (Sugiura) [PDF]
  • LLM-jp-3 VILA: Construction of Japanese Multimodal Data and Powerful Japanese Multimodal Model (Sasagawa) [PDF]

<Model Building WG>

  • Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization (Nakamura) [PDF]

<Safety WG>

  • Large-Scale Human Evaluation of LLM Safety (Takahashi) [PDF]
  • AnswerCarefully: AnswerCarefully: A Dataset for Promoting Safety of Japanese LLMs (Suzuki)[PDF]
  • Developing a Dataset of Misinformation from Social Media and an Accuracy Benchmark for Large Language Models (Nakazato)[PDF]
  • Development of Prompt Attack Data Collection Application for LLMs and Analysis of Collected Data Characteristics (Hayashi)[PDF]

<Corpus Construction WG>

  • A Comprehensive Analysis of Memorization in Large Language Models (Kiyomaru) [PDF]
  • Detection of Sensitive Personal Information in the Pre-training Corpus for Large Language Models (Minamoto) [PDF]
  • Integrated Framework for LLM Domain Adaptation Based on Synthetic Data (Ogawa) [PDF]

2025-2-25

  • LLM-jp Status Report (Kurohashi) Oral
  • Real-world Interaction Working Group (Ogata) [PDF]
  • Safety Working Group (Sekine) [PDF]
  • Japan AI Safety Institute (Semitsu) [PDF]
  • Model Building Working Group (Suzuki) [PDF]
  • Multi-modal Working Group (Okazaki) [PDF]
  • Corpus Construction Working Group (Kawahara) [PDF]
  • Evaluation and Tuning Working Group (Miyao) [PDF]
  • Dataflow Architecture Achieving 198 Tokens per Second with DeepSeek R1 671B (Kenichi Hayashi/SambaNova System) [PDF]
  • PLaMo 2 Tokenizer: Keys to Token Efficiency (Kentaro Imajo/Preferred Networks) [PDF]
  • Training progress of LLM-jp-3 models: analysis on downstream performance (Nishida/Oda) [PDF]

2025-1-16

  • LLM-jp Status Report (Kurohashi) [PDF]
  • Model Building Working Group (Suzuki) [PDF]
  • Multi-modal Working Group (Okazaki) [PDF]
  • Evaluation and Tuning Working Group (Miyao) [PDF]
  • Real-world Interaction Working Group (Ogata) [PDF]
  • Corpus Construction Working Group (Kawahara) [PDF]
  • Junichiro Takahashi/The University of Tokyo [PDF]
  • Kaito Baba/The University of Tokyo [PDF]
  • Efforts to Efficiently Create High-Quality LLM Datasets (Yujiro Terazawa/APTO,Inc) [PDF]
  • LLM Training Using Synthetic Data (Kiyomaru) [PDF]

2024-11-26

  • LLM-jp Status Report (Kurohashi) Oral Presenation
  • Development and Evaluation of Tanuki (Kan Hatakeyama/Institute of Science Tokyo) [PDF]
  • EMNLP2024 Report (Takagi) [PDF] (Kodama) [PDF] (Liu) [PDF]
  • Safety Working Group (Sekine) [PDF]
  • Corpus Construction Working Group (Kawahara) [PDF]
  • Evaluation and Tuning Working Group (Miyao) [PDF]
  • Real-world Interaction Working Group (Ogata) [PDF]
  • Model Building Working Group (Suzuki) [PDF]
  • Multi-modal Working Group (Okazaki) [PDF]
  • Discussion on the Training Progress of LLM-jp-3 172B (Oda) [PDF]

2024-10-29

2024-08-27

2024-07-30

2024-06-25

2024-05-26

2024-03-26

2024-01-22

2023-11-29

2023-10-18

2023-09-04

2023-07-20

2023-06-19

2023-05-15