AT THE INTERSECTION OF SUPERCOMPUTING AND DEEP LEARNING
Our research group is working on various applications of large-scale distributed parallelism using supercomputers. In recent years, deep learning has been the focal point of attention in image recognition, natural language processing, and reinforcement learning, etc. The scale of deep neural nets used in these fields is increasing exponentially, and training these networks is becoming impossible without the use of supercomputers. However, simply running existing deep learning frameworks on supercomputers will not immediately improve the speed of the training nor the accuracy of the resulting model. It is necessary to solve the issues specific to large-scale distributed training one by one before we can investigate the scaling laws of deep neural networks. Scientific computing, which have been performed on supercomputers for a long time, also require continuous research on algorithms and implementation methods to match the ever-changing computer architectures. Furthermore, since the performance of computers continues to improve exponentially due to Moore's Law, the calculations performed on today's supercomputers will be able to be performed on a local desktop computer in 10 years. In other words, solving the problems on today's supercomputers is equivalent to solving the research problems of 10 years from now in advance.
COMPUTATIONAL RESOURCES
In our laboratory, we have access to some of the largest supercomputers in Japan, including TSUBAME at Tokyo Institute of Technology, Miyabi at the University of Tokyo, ABCI at AIST, and Fugaku at RIKEN. In addition, by actively using the Grand Challenge System, which allows us to have exclusive access to the entire system of these supercomputers, we are able to use one of the largest amounts of computing resources among academic research groups. In addition, by concluding collaborative research agreements (MOUs) [https://adac.ornl.gov] with 14 major supercomputer centers around the world, we have access to the world's largest supercomputers such as Frontier at ORNL, Aurora at ANL, LUMI at CSC, and Alps at ETH/CSCS. Computations that would take weeks in a normal research groups’ computing environment can be performed in a few hours in our group.
JOINT RESEARCH PROJECTS
Our expertise in large-scale computation on supercomputers and our vast amount of computing resources are useful in many research fields, including deep learning and scientific computing. Currently, we are participating in many joint research projects with domestic and international research institutions and companies, both within and outside the university. Within our university, we are collaborating with Okazaki Group on the Japanese large language model Swallow, and with Shinoda, Inoue, and Sato Groups on computer vision. Externally, we are collaborating with the Khan Group at RIKEN AIP on Bayesian deep learning, with the AI Research Center at AIST on vision-language models, and with NII on the Japanese large language model llm-jp. Outside Japan, we are collaborating with the top supercomputer centers such as ORNL, ANL, LLNL, CSC, and ETH. This means that you can choose from a wide range of research topics, or if you want to find a new research topic on your own, you are not limited to the expertise of your supervisor alone, but can receive appropriate support through our collaborators.