화학공학소재연구정보센터
학회 한국화학공학회
학술대회 2022년 봄 (04/20 ~ 04/23, 제주국제컨벤션센터)
권호 28권 1호, p.121
발표분야 [주제 2] 기계학습
제목 N-step lookahead reinforcement learning based Bayesian optimization for limited experiment budget
초록 Bayesian optimization(BO) is a sequential decision-making strategy for efficiently finding a global optimum for black-box optimization problems. In chemical engineering, where data generation through experiments is expensive, decisions based on BO are crucial. However, widely used standard BOs are one-step optimal. This means conventional BO only considers the immediate improvement. However, none of the real-world problems can be solved in a single step of iteration. Therefore, a strategy of making decisions while looking N-step ahead is required. This work suggests Reinforcement learning(RL) based BO to achieve this goal. Sequential experiments can be viewed as stochastic dynamic programming(SDP) problems, and RL is the method to solve SDP in the most near-optimum way. In this research, the experiment horizon was assumed to be fixed and the efficiency of optimization algorithms were compared on benchmark functions. Once the number of experiments are assumed to be limited, the suggested RL based BO has a shrinking horizon for its lookahead decisions. It is revealed that suggested RL based BO has superior data efficiency compared to EI based BO and other conventional methods.
저자 천무진, 변하은, 이재형
소속 한국과학기술원
키워드 공정시스템(Process Systems Engineering)
E-Mail
원문파일 초록 보기