Home / Papers / RL-OPC: Mask Optimization With Deep Reinforcement Learning

RL-OPC: Mask Optimization With Deep Reinforcement Learning

6 Citations2024
Xiaoxiao Liang, Yikang Ouyang, Haoyu Yang
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems

This article pioneer introducing the reinforcement learning (RL) model for mask optimization, which directly optimizes the preferred objective without leveraging a differentiable proxy, and outperforms state-of-the-art solutions.

Abstract

Mask optimization is a vital step in the VLSI manufacturing process in advanced technology nodes. As one of the most representative techniques, optical proximity correction (OPC) is widely applied to enhance printability. Since conventional OPC methods consume prohibitive computational overhead, recent research has applied machine learning techniques for efficient mask optimization. However, existing discriminative learning models rely on a given dataset for supervised training, and generative learning models usually leverage a proxy optimization objective for end-to-end learning, which may limit the feasibility. In this article, we pioneer introducing the reinforcement learning (RL) model for mask optimization, which directly optimizes the preferred objective without leveraging a differentiable proxy. Intensive experiments show that our method outperforms state-of-the-art solutions, including academic approaches and commercial toolkits.