Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (3): 697-708.DOI: 10.11996/JG.j.2095-302X.2025030697

• Digital Design and Manufacture • Previous Articles     Next Articles

Transport-and-packing with buffer via deep reinforcement learning

LEI Yulin(), LIU Ligang()   

  1. School of Mathematical Sciences, University of Science and Technology of China, Hefei Anhui 230026, China
  • Received:2024-10-19 Accepted:2025-02-19 Online:2025-06-30 Published:2025-06-13
  • Contact: LIU Ligang
  • About author:First author contact:

    LEI Yulin (2000-), master student. His main research interests cover computer graphics and deep learning. E-mail:yllei@mail.ustc.edu.cn

  • Supported by:
    National Natural Science Foundation of China(62025207)

Abstract:

Addressing the challenge of limited container space utilization caused by initial object stacking constraints in physical scenarios, a neural optimization model based on a deep reinforcement learning framework was proposed for bufferable object transportation and packing, incorporating a buffer transfer mechanism to enhance container packing efficiency. The state encoder dynamically encoded priority information extracted from a priority graph and buffer information, effectively managed object stacking relationships, and leveraged the transfer capacity of the buffer zone. The sequence decoder perceived the current container state and employed an attention mechanism to calculate selection probabilities for candidate rotation state sequences, adaptively selecting sequences for either transfer or packing. Subsequently, the target decoder took the geometric and buffer information of the selected states as input, integrated the accumulated information from the sequence decoder to construct a conditional query vector, and performed attention aggregation on the encoded feature vectors to efficiently decide whether to buffer or pack objects. The REINFORCE algorithm with a baseline was employed to train the network, yielding optimized strategies for bufferable object packing. Experimental results on 2D and 3D RAND datasets demonstrated an approximate 4% improvement in container packing utilization compared to the advanced TAP-Net model, significantly outperforming heuristic methods designed for this newly defined problem. Furthermore, models trained on a fixed number of objects effectively generalized to packing instances involving a larger number of objects.

Key words: bin packing problem, deep reinforcement learning, neural optimization, combinatorial optimization, attention mechanism

CLC Number: