TY - GEN
T1 - Exploratory Combinatorial Optimization Problem Solving via Gauge Transformation
AU - Pu, Tianle
AU - Fan, Changjun
AU - Shen, Mutian
AU - Lu, Yizhou
AU - Zeng, Li
AU - Nussinov, Zohar
AU - Chen, Chao
AU - Liu, Zhong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The combinatorial optimization problems (COPs) over graph are of great significance both in theory and practice, covering a wide range of scenarios in daily life and industrial production. Recent years, reinforcement learning (RL) based models have emerged as a promising direction, which treat solving the COPs as a heuristic learning problem. However, current finite-horizon Markov Decision Process (MDP) based RL models are not allowed to explore adquately for improving solutions at test time, which may be necessary given the complexity of NP-hard optimization tasks. Some recent attempts solve this issue by focusing on reward design and state feature engineering, which are tedious and ad-hoc. To address this challenge, we introduce a physics-inspired technique called gauge transformation (GT), which is highly effective in enabling RL agents to explore and continuously enhance solution quality during testing. GT seamlessly transforms any state within the MDP back to its initial state, allowing the RL agent to continue exploration within the transformed space. Empirically, we demonstrate that traditional RL models equipped with the GT technique achieve the SOTA performance on the MaxCut problem. Moreover, GT is exclusively applied during testing and does not alter the training phase of the model. It can be readily integrated into existing RL models, providing a pathway for more effective exploration in solving the COPs.
AB - The combinatorial optimization problems (COPs) over graph are of great significance both in theory and practice, covering a wide range of scenarios in daily life and industrial production. Recent years, reinforcement learning (RL) based models have emerged as a promising direction, which treat solving the COPs as a heuristic learning problem. However, current finite-horizon Markov Decision Process (MDP) based RL models are not allowed to explore adquately for improving solutions at test time, which may be necessary given the complexity of NP-hard optimization tasks. Some recent attempts solve this issue by focusing on reward design and state feature engineering, which are tedious and ad-hoc. To address this challenge, we introduce a physics-inspired technique called gauge transformation (GT), which is highly effective in enabling RL agents to explore and continuously enhance solution quality during testing. GT seamlessly transforms any state within the MDP back to its initial state, allowing the RL agent to continue exploration within the transformed space. Empirically, we demonstrate that traditional RL models equipped with the GT technique achieve the SOTA performance on the MaxCut problem. Moreover, GT is exclusively applied during testing and does not alter the training phase of the model. It can be readily integrated into existing RL models, providing a pathway for more effective exploration in solving the COPs.
KW - Combinatorial Optimization
KW - Gauge Transformation
KW - Graph Neural Network
KW - MaxCut Problem
KW - Reinforcement Learning
UR - https://www.scopus.com/pages/publications/86000227789
U2 - 10.1109/ICDM59182.2024.00099
DO - 10.1109/ICDM59182.2024.00099
M3 - Conference contribution
AN - SCOPUS:86000227789
T3 - Proceedings - IEEE International Conference on Data Mining, ICDM
SP - 821
EP - 826
BT - Proceedings - 24th IEEE International Conference on Data Mining, ICDM 2024
A2 - Baralis, Elena
A2 - Zhang, Kun
A2 - Damiani, Ernesto
A2 - Debbah, Merouane
A2 - Kalnis, Panos
A2 - Wu, Xindong
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 24th IEEE International Conference on Data Mining, ICDM 2024
Y2 - 9 December 2024 through 12 December 2024
ER -