RIKEN Center for Computational Science High Performance Artificial Intelligence Systems Research Team
Team Leader: Mohamed Wahib (Ph.D.)
Research Summary
The High Performance Artificial Intelligence Systems Research Team is an R-CCS laboratory focusing on convergence of HPC and AI, namely high performance systems, software, and algorithms research for artificial intelligence/machine learning. In collaboration with other research institutes in HPC and AI-related research in Japan as well as globally, it seeks to develop next-generation AI technology that will utilize state-of-the-art high-performance computation facilities, including Fugaku. Specifically, we conduct research on next-generation AI systems by focusing on the following topics:
- 1.Extreme speedup and scalability of deep learning:
Achieve extreme scalability of deep learning in large-scale supercomputing environments including the post-K, extending the latest algorithms and frameworks for deep learning. - 2.Performance analysis of deep learning:
Accelerate computational kernels for AI over the state-of-the-art hardware architectures by analyzing algorithms for deep learning and other machine learning/AI, measuring their performance and constructing their performance models. - 3.Acceleration of modern AI algorithms:
Accelerate advanced AI algorithms, such as ultra-deep neural networks and high-resolution GAN over images, those that require massive computational resources, using extreme-scale deep learning systems. - 4.Science Enabled by AI:
Extend the capability of AI to qualitatively and quantitatively advance scientific and engineering. - 5.Intelligent programming systems:
Use AI to auto-generate programs that can adapt to and withstand the complexity and divergence of hardware design.
Main Research Fields
- Informatics
Related Research Fields
- High Performance Computing
- Parallel Distributed Processing
- Computer Architecture
Keywords
- High Performance Artificial Intelligence Systems
- Intelligent Programming Systems
- Performance Modeling of AI Systems e.g. Deep Learning
- Scalable Deep Learning
- Convergence of AI and Simulation
Selected Publications
Papers with an asterisk(*) are based on research conducted outside of RIKEN.
- 1.
*Thao Nguyen Truong, Balazs Gerofi, Edgar Josafat Martinez-Noriega, Francois Trahay, Mohamed Wahib, KAKURENBO:
"Adaptively Hiding Samples in Deep Neural Network Training",
Advances in Neural Information Processing Systems 2023 (NeurIPS 2023) - 2.
*Huaipeng Zhang, Nhut-Minh Ho, Yigit Polat Dogukan, Peng Chen, Mohamed Wahib, Truong Thao Nguyen, Jintao Meng, Rick Siow Mong Goh, Satoshi Matsuoka, Tao Luo, Weng-Fai Wong,
"Simeuro: A Hybrid CPU-GPU Parallel Simulator for Neuromorphic Computing Chips,"
in IEEE Transactions on Parallel and Distributed Systems, vol. 34, no. 10, pp. 2767-2782, Oct. 2023 - 3.
*Lingqi, Mohamed Wahib, Chen Peng, Jintao Meng, Xiao Wang, Toshio Endo, Satoshi Matsuoka, PERKS:
"a Locality-Optimized Execution Model for Iterative Memory-bound GPU Applications",
ACM 37th International Confernce on Supercomputing (ACM ICS 2023) - 4.
*Jintao Meng, Chen Zhuang, Peng Chen, Mohamed Wahib, Bertil Schmidt, Xiao Wang, Haidong Lan, Dou Wu, Minwen Deng, Yanjie Wei, Shenzhong Feng,
"Automatic Generation of High-Performance Convolution Kernels on ARM CPUs for Deep Learning"
in IEEE Transactions on Parallel & Distributed Systems, vol. 34, April 2022. - 5.
*Jintao Meng, Peng Chen, Mingjun Yang, Mohamed Wahib, Yanjie Wei, Shengzhong Feng, Wei Liu, Junzhou Huang,
"Boosting the Predictive Performance with Aqueous Solubility Dataset Curation",
Nature Scientific Data, March 2022 - 6.
*Albert Khaira, Truong Thao Nguyen, Leonardo Bautista Gomez, Ryousei Takano, Rosa Badia, Mohamed Wahib:
"An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of Convolutional Neural Networks",
30th ACM International Symposium on High-Performance Parallel and Distributed Computing (HPDC 2021) - 7.
Peng Chen, Mohamed Wahib, Xiao Wang, Takahiro Hirofuchi, Hirotaka Ogawa, Ander Biguri, Richard Boardman, Thomas Blumensath, Satoshi Matsuoka,
"Scalable FBP Decomposition for Cone-Beam CT Reconstruction",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2021) - 8.
Fareed Mohammad Qararyah, Mohamed Wahib, Doga Dikbayır, Mehmet Esat Belviranl, Didem Unat,
"A computational-graph Partitioning Method for Training Memory-constrained DNNs",
Elsevier Parallel Computing, Volume 104 pp. 102-117, July 2021 - 9.
Mohamed Wahib, Haoyu Zhang, Truong Thao Nguyen, Aleksandr Drozd, Jens Domke, Lingqi Zhang, Ryousei Takano, Satoshi Matsuoka:
"Scaling Deep Learning Workloads Beyond Memory Capacity",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2020) - 10.
Chen Peng,Wahib Mohamed,Takizawa Shinichiro,Matsuoka Satoshi:
"A Versatile Software Systolic Execution Model for GPU Memory Bound Kernels",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2019)
Related Links
Lab Members
Principal investigator
- Mohamed Wahib
- Team Leader
Core members
- Jun Igarashi
- Senior Scientist
- Aleksandr Drozd
- Research Scientist
- Emil Vatai
- Research Scientist
- Zhengyang Bai
- Postdoctoral Researcher
- Lingqi Zhang
- Postdoctoral Researcher
- Joao Eduardo Batista
- Postdoctoral Researcher
- Du Wu
- Junior Research Associate
- Chen Zhuang
- Junior Research Associate
- Balazs Gerofi
- Visiting Scientist
- Rio Yokota
- Visiting Scientist
Careers
Position | Deadline |
---|---|
Seeking a few Research Scientists or Postdoctoral Researchers (R-CCS2202) | Open until filled |
Contact Information
Nihonbashi 1-chome Mitsui Building, 15th floor,
1-4-1 Nihonbashi,
Chuo-ku, Tokyo
103-0027, Japan
Email: mohamed.attia [at] riken.jp