A K-means Supported Reinforcement Learning Framework to Multi-dimensional Knapsack

dc.contributor.authorBushaj, Sabahen
dc.contributor.authorBüyüktahtakın, İ. Esraen
dc.date.accessioned2025-03-14T13:56:30Zen
dc.date.available2025-03-14T13:56:30Zen
dc.date.issued2024-02-15en
dc.description.abstractIn this paper, we address the difficulty of solving large-scale multi-dimensional knapsack instances (MKP), presenting a novel deep reinforcement learning (DRL) framework. In this DRL framework, we train different agents compatible with a discrete action space for sequential decision-making while still satisfying any resource constraint of the MKP. This novel framework incorporates the decision variable values in the 2D DRL where the agent is responsible for assigning a value of 1 or 0 to each of the variables. To the best of our knowledge, this is the first DRL model of its kind in which a 2D environment is formulated, and an element of the DRL solution matrix represents an item of the MKP. Our framework is configured to solve MKP instances of different dimensions and distributions. We propose a K-means approach to obtain an initial feasible solution that is used to train the DRL agent. We train four different agents in our framework and present the results comparing each of them with the CPLEX commercial solver. The results show that our agents can learn and generalize over instances with different sizes and distributions. Our DRL framework shows that it can solve medium-sized instances at least 45 times faster in CPU solution time and at least 10 times faster for large instances, with a maximum solution gap of 0.28% compared to the performance of CPLEX. Furthermore, at least 95% of the items are predicted in line with the CPLEX solution. Computations with DRL also provide a better optimality gap with respect to state-of-the-art approaches.en
dc.description.versionPublished versionen
dc.format.extentPages 655-685en
dc.format.extent31 page(s)en
dc.format.mimetypeapplication/pdfen
dc.identifier.doihttps://doi.org/10.1007/s10898-024-01364-6en
dc.identifier.eissn1573-2916en
dc.identifier.issn0925-5001en
dc.identifier.issue3en
dc.identifier.orcidBuyuktahtakin Toy, Esra [0000-0001-8928-2638]en
dc.identifier.urihttps://hdl.handle.net/10919/124864en
dc.identifier.volume89en
dc.language.isoenen
dc.publisherSpringeren
dc.rightsCreative Commons Attribution 4.0 Internationalen
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/en
dc.subjectMulti-dimensional Knapsack problemen
dc.subjectK-meansen
dc.subjectDeep reinforcement learningen
dc.subjectCombinatorial optimizationen
dc.subjectMixed integer programmingen
dc.subjectHeuristicsen
dc.titleA K-means Supported Reinforcement Learning Framework to Multi-dimensional Knapsacken
dc.title.serialJournal of Global Optimizationen
dc.typeArticle - Refereeden
dc.type.dcmitypeTexten
dc.type.otherArticleen
dc.type.otherJournalen
pubs.organisational-groupVirginia Techen
pubs.organisational-groupVirginia Tech/Engineeringen
pubs.organisational-groupVirginia Tech/Engineering/Industrial and Systems Engineeringen
pubs.organisational-groupVirginia Tech/All T&R Facultyen
pubs.organisational-groupVirginia Tech/Engineering/COE T&R Facultyen

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
BushajBuyuktahtakin2024_K-Means.pdf
Size:
2.57 MB
Format:
Adobe Portable Document Format
Description:
Published version
License bundle
Now showing 1 - 1 of 1
Name:
license.txt
Size:
1.5 KB
Format:
Plain Text
Description: