TY - CHAP
T1 - High-Performance Computing Environment with Cooperation Between Supercomputer and Cloud
AU - Kotani, Toshihiro
AU - Gotoh, Yusuke
N1 - Funding Information:
This work was supported by JSPS KAKENHI Grant Number 18K11265.
Publisher Copyright:
© Springer Nature Switzerland AG 2020.
PY - 2020
Y1 - 2020
N2 - Due to the recent popularization of machine learning, such a deep reinforcement learning as AlphaGO has advanced to analyze large-scale data and is attracting great attention. In deep reinforcement learning, users evaluate many functions in large-scale computer environments, including supercomputer and cloud systems. Cloud services can provide computer resources based on the scale of the computer environment desired by users. On the other hand, in conventional large-scale computer environment that only consists of CPUs or GPUs, the processing time greatly increases according to the scale of the calculation processing. In this paper, we propose a high-performance computing environment for deep reinforcement learning that links supercomputer and cloud systems. Our proposed system can construct a high-performance computing environment based on the scale of the computing process by the cooperation of the supercomputing and cloud systems with short physical distance and short network distance. In our evaluation of deep reinforcement learning using our proposed system, we confirmed that computer resources can be effectively used by allocating suitable processing for the supercomputer and the cloud according to the usage situations of the CPU, the GPU, and the memory.
AB - Due to the recent popularization of machine learning, such a deep reinforcement learning as AlphaGO has advanced to analyze large-scale data and is attracting great attention. In deep reinforcement learning, users evaluate many functions in large-scale computer environments, including supercomputer and cloud systems. Cloud services can provide computer resources based on the scale of the computer environment desired by users. On the other hand, in conventional large-scale computer environment that only consists of CPUs or GPUs, the processing time greatly increases according to the scale of the calculation processing. In this paper, we propose a high-performance computing environment for deep reinforcement learning that links supercomputer and cloud systems. Our proposed system can construct a high-performance computing environment based on the scale of the computing process by the cooperation of the supercomputing and cloud systems with short physical distance and short network distance. In our evaluation of deep reinforcement learning using our proposed system, we confirmed that computer resources can be effectively used by allocating suitable processing for the supercomputer and the cloud according to the usage situations of the CPU, the GPU, and the memory.
UR - http://www.scopus.com/inward/record.url?scp=85074686601&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85074686601&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-33509-0_40
DO - 10.1007/978-3-030-33509-0_40
M3 - Chapter
AN - SCOPUS:85074686601
T3 - Lecture Notes in Networks and Systems
SP - 433
EP - 443
BT - Lecture Notes in Networks and Systems
PB - Springer
ER -