Skip to main content
construction release_alert
Scholars@Duke will be down for maintenance for approximately one hour starting Tuesday, 11/11 @1pm ET
cancel
Journal cover image

Deep reinforcement learning for solving the stochastic e-waste collection problem

Publication ,  Journal Article
Nguyen, DVA; Gunawan, A; Misir, M; Hui, LK; Vansteenwegen, P
Published in: European Journal of Operational Research
November 16, 2025

With the growing influence of the internet and information technology, Electrical and Electronic Equipment (EEE) has become a gateway to technological innovations. However, discarded devices, also called e-waste, pose a significant threat to the environment and human health if not properly treated, disposed of, or recycled. In this study, we extend a novel model for the e-waste collection in an urban context: the Heterogeneous VRP with Multiple Time Windows and Stochastic Travel Times (HVRP-MTWSTT). We propose a solution method that employs deep reinforcement learning to guide local search heuristics (DRL-LSH). The contributions of this paper are as follows: (1) HVRP-MTWSTT represents the first stochastic VRP in the context of the e-waste collection problem, incorporating complex constraints such as multiple time windows across a multi-period horizon with a heterogeneous vehicle fleet, (2) The DRL-LSH model uses deep reinforcement learning to provide an online adaptive operator selection layer, selecting the appropriate heuristic based on the search state. The computational experiments demonstrate that DRL-LSH outperforms the state-of-the-art hyperheuristic method by 24.26% on large-scale benchmark instances, with the performance gap increasing as the problem size grows. Additionally, to demonstrate the capability of DRL-LSH in addressing real-world problems, we tested and compared it with reference metaheuristic and hyperheuristic algorithms using a real-world e-waste collection case study in Singapore. The results showed that DRL-LSH significantly outperformed the reference algorithms on a real-world instance in terms of operating profit.

Duke Scholars

Published In

European Journal of Operational Research

DOI

ISSN

0377-2217

Publication Date

November 16, 2025

Volume

327

Issue

1

Start / End Page

309 / 325

Related Subject Headings

  • Operations Research
  • 49 Mathematical sciences
  • 46 Information and computing sciences
  • 40 Engineering
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Nguyen, D. V. A., Gunawan, A., Misir, M., Hui, L. K., & Vansteenwegen, P. (2025). Deep reinforcement learning for solving the stochastic e-waste collection problem (Accepted). European Journal of Operational Research, 327(1), 309–325. https://doi.org/10.1016/j.ejor.2025.04.033
Nguyen, D. V. A., A. Gunawan, M. Misir, L. K. Hui, and P. Vansteenwegen. “Deep reinforcement learning for solving the stochastic e-waste collection problem (Accepted).” European Journal of Operational Research 327, no. 1 (November 16, 2025): 309–25. https://doi.org/10.1016/j.ejor.2025.04.033.
Nguyen DVA, Gunawan A, Misir M, Hui LK, Vansteenwegen P. Deep reinforcement learning for solving the stochastic e-waste collection problem (Accepted). European Journal of Operational Research. 2025 Nov 16;327(1):309–25.
Nguyen, D. V. A., et al. “Deep reinforcement learning for solving the stochastic e-waste collection problem (Accepted).” European Journal of Operational Research, vol. 327, no. 1, Nov. 2025, pp. 309–25. Scopus, doi:10.1016/j.ejor.2025.04.033.
Nguyen DVA, Gunawan A, Misir M, Hui LK, Vansteenwegen P. Deep reinforcement learning for solving the stochastic e-waste collection problem (Accepted). European Journal of Operational Research. 2025 Nov 16;327(1):309–325.
Journal cover image

Published In

European Journal of Operational Research

DOI

ISSN

0377-2217

Publication Date

November 16, 2025

Volume

327

Issue

1

Start / End Page

309 / 325

Related Subject Headings

  • Operations Research
  • 49 Mathematical sciences
  • 46 Information and computing sciences
  • 40 Engineering