Improved Proximal Policy Optimization Algorithm for Multi-objective Disassembly Line Balancing Considering Hazardous Tasks
Document Type
Conference Proceeding
Publication Date
1-1-2024
Abstract
As disassembly lines evolve, keeping workers safe from all potential dangers has become paramount. This work proposes an improved approach, Proximal Policy Optimization for Disassembly Line(PPO-DL), to address Multi-objective Dis-assembly Line Balancing Problem(MDLBP) that integrates task switching time and hazardous tasks. The MDLBP model aims to optimize disassembly net profit, balance workstation utilization, and mitigate penalties from hazardous tasks. The proposed PPO-DL modifies the action space of PPO to accommodate MD LBP characteristics, ensuring efficient learning by restricting actions to valid ranges through action masking. Comparative experiments with PPO, Q-learning, and A2C demonstrate PPO-DL's superiority in finding near-optimal solutions across various scenarios.
Identifier
85213360599 (Scopus)
ISBN
[9798350365221]
Publication Title
ICNSC 2024 - 21st International Conference on Networking, Sensing and Control: Artificial Intelligence for the Next Industrial Revolution
External Full Text Location
https://doi.org/10.1109/ICNSC62968.2024.10760016
Recommended Citation
Guo, Zeyu; Guo, Xiwang; Wang, Jiacun; Qin, Shujin; Qi, Liang; and Kang, Qi, "Improved Proximal Policy Optimization Algorithm for Multi-objective Disassembly Line Balancing Considering Hazardous Tasks" (2024). Faculty Publications. 758.
https://digitalcommons.njit.edu/fac_pubs/758