1. Home
  2. Research
  3. Centers & Labs
  4. RIKEN Center for Computational Science

RIKEN Center for Computational Science High Performance Artificial Intelligence Systems Research Team

Team Leader: Mohamed Wahib (Ph.D.)

Research Summary

Mohamed Wahib(Ph.D.)

The High Performance Artificial Intelligence Systems Research Team is an R-CCS laboratory focusing on convergence of HPC and AI, namely high performance systems, software, and algorithms research for artificial intelligence/machine learning. In collaboration with other research institutes in HPC and AI-related research in Japan as well as globally, it seeks to develop next-generation AI technology that will utilize state-of-the-art high-performance computation facilities, including Fugaku. Specifically, we conduct research on next-generation AI systems by focusing on the following topics:

  • 1.Extreme speedup and scalability of deep learning:
    Achieve extreme scalability of deep learning in large-scale supercomputing environments including the post-K, extending the latest algorithms and frameworks for deep learning.
  • 2.Performance analysis of deep learning:
    Accelerate computational kernels for AI over the state-of-the-art hardware architectures by analyzing algorithms for deep learning and other machine learning/AI, measuring their performance and constructing their performance models.
  • 3.Acceleration of modern AI algorithms:
    Accelerate advanced AI algorithms, such as ultra-deep neural networks and high-resolution GAN over images, those that require massive computational resources, using extreme-scale deep learning systems.
  • 4.Science Enabled by AI:
    Extend the capability of AI to qualitatively and quantitatively advance scientific and engineering.
  • 5.Intelligent programming systems:
    Use AI to auto-generate programs that can adapt to and withstand the complexity and divergence of hardware design.

Main Research Fields

  • Informatics

Related Research Fields

  • High Performance Computing
  • Parallel Distributed Processing
  • Computer Architecture


  • High Performance Artificial Intelligence Systems
  • Intelligent Programming Systems
  • Performance Modeling of AI Systems e.g. Deep Learning
  • Scalable Deep Learning
  • Convergence of AI and Simulation

Selected Publications

Papers with an asterisk(*) are based on research conducted outside of RIKEN.

  • 1. *Thao Nguyen Truong, Balazs Gerofi, Edgar Josafat Martinez-Noriega, Francois Trahay, Mohamed Wahib, KAKURENBO:
    "Adaptively Hiding Samples in Deep Neural Network Training",
    Advances in Neural Information Processing Systems 2023 (NeurIPS 2023)
  • 2. *Huaipeng Zhang, Nhut-Minh Ho, Yigit Polat Dogukan, Peng Chen, Mohamed Wahib, Truong Thao Nguyen, Jintao Meng, Rick Siow Mong Goh, Satoshi Matsuoka, Tao Luo, Weng-Fai Wong,
    "Simeuro: A Hybrid CPU-GPU Parallel Simulator for Neuromorphic Computing Chips,"
    in IEEE Transactions on Parallel and Distributed Systems, vol. 34, no. 10, pp. 2767-2782, Oct. 2023
  • 3. *Lingqi, Mohamed Wahib, Chen Peng, Jintao Meng, Xiao Wang, Toshio Endo, Satoshi Matsuoka, PERKS:
    "a Locality-Optimized Execution Model for Iterative Memory-bound GPU Applications",
    ACM 37th International Confernce on Supercomputing (ACM ICS 2023)
  • 4. *Jintao Meng, Chen Zhuang, Peng Chen, Mohamed Wahib, Bertil Schmidt, Xiao Wang, Haidong Lan, Dou Wu, Minwen Deng, Yanjie Wei, Shenzhong Feng,
    "Automatic Generation of High-Performance Convolution Kernels on ARM CPUs for Deep Learning"
    in IEEE Transactions on Parallel & Distributed Systems, vol. 34, April 2022.
  • 5. *Jintao Meng, Peng Chen, Mingjun Yang, Mohamed Wahib, Yanjie Wei, Shengzhong Feng, Wei Liu, Junzhou Huang,
    "Boosting the Predictive Performance with Aqueous Solubility Dataset Curation",
    Nature Scientific Data, March 2022
  • 6. *Albert Khaira, Truong Thao Nguyen, Leonardo Bautista Gomez, Ryousei Takano, Rosa Badia, Mohamed Wahib:
    "An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of Convolutional Neural Networks",
    30th ACM International Symposium on High-Performance Parallel and Distributed Computing (HPDC 2021)
  • 7. Peng Chen, Mohamed Wahib, Xiao Wang, Takahiro Hirofuchi, Hirotaka Ogawa, Ander Biguri, Richard Boardman, Thomas Blumensath, Satoshi Matsuoka,
    "Scalable FBP Decomposition for Cone-Beam CT Reconstruction",
    International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2021)
  • 8. Fareed Mohammad Qararyah, Mohamed Wahib, Doga Dikbayır, Mehmet Esat Belviranl, Didem Unat,
    "A computational-graph Partitioning Method for Training Memory-constrained DNNs",
    Elsevier Parallel Computing, Volume 104 pp. 102-117, July 2021
  • 9. Mohamed Wahib, Haoyu Zhang, Truong Thao Nguyen, Aleksandr Drozd, Jens Domke, Lingqi Zhang, Ryousei Takano, Satoshi Matsuoka:
    "Scaling Deep Learning Workloads Beyond Memory Capacity",
    International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2020)
  • 10. Chen Peng,Wahib Mohamed,Takizawa Shinichiro,Matsuoka Satoshi:
    "A Versatile Software Systolic Execution Model for GPU Memory Bound Kernels",
    International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2019)

Related Links

Lab Members

Principal investigator

Mohamed Wahib
Team Leader

Core members

Jun Igarashi
Senior Scientist
Aleksandr Drozd
Research Scientist
Emil Vatai
Research Scientist
Zhengyang Bai
Postdoctoral Researcher
Lingqi Zhang
Postdoctoral Researcher
Joao Eduardo Batista
Postdoctoral Researcher
Du Wu
Junior Research Associate
Chen Zhuang
Junior Research Associate
Balazs Gerofi
Visiting Scientist
Rio Yokota
Visiting Scientist


Position Deadline
Seeking a few Research Scientists or Postdoctoral Researchers (R-CCS2202) Open until filled

Contact Information

Nihonbashi 1-chome Mitsui Building, 15th floor,
1-4-1 Nihonbashi,
Chuo-ku, Tokyo
103-0027, Japan
Email: mohamed.attia [at] riken.jp