Please use this identifier to cite or link to this item:
Title: A biased-randomized learnheuristic for solving the team orienteering problem with dynamic rewards
Author: Reyes Rubiano, Lorena Silvana
Juan Pérez, Ángel Alejandro
Bayliss, Christopher
Panadero Martínez, Javier
Faulin Fajardo, Francisco Javier
Copado Méndez, Pedro Jesús
Others: Universidad Pública de Navarra
Universitat Oberta de Catalunya. Internet Interdisciplinary Institute (IN3)
Keywords: transportation
team orienteering problem
dynamic inputs
biased randomization
Issue Date: 25-Apr-2020
Publisher: Transportation Research Procedia
Citation: Reyes-Rubiano, L., Juan, A.A., Bayliss, C., Panadero, J., Faulin, J. & Copado, P. (2020). A biased-randomized learnheuristic for solving the team orienteering problem with dynamic rewards. Transportation Research Procedia, 47(), 680-687. doi: 10.1016/j.trpro.2020.03.147
Published in: Euro Working Group on Transportation Meeting (EWGT), Barcelona, 18-20 de setembre de 2019
Also see:
Abstract: In this paper we discuss the team orienteering problem (TOP) with dynamic inputs. In the static version of the TOP, a fixed reward is obtained after visiting each node. Hence, given a limited fleet of vehicles and a threshold time, the goal is to design the set of routes that maximize the total reward collected. While this static version can be efficiently tackled using a biased-randomized heuristic (BR-H), dealing with the dynamic version requires extending the BR-H into a learnheuristic (BR-LH). With that purpose, a 'learning' (white-box) mechanism is incorporated to the heuristic in order to consider the variations in the observed rewards, which follow an unknown (black-box) pattern. In particular, we assume that: (i) each node in the network has a 'base' or standard reward value; and (ii) depending on the node's position inside its route, the actual reward value might differ from the base one according to the aforementioned unknown pattern. As new observations of this black-box pattern are obtained, the white-box mechanism generates better estimates for the actual rewards after each new decision. Accordingly, better solutions can be generated by using this predictive mechanism. Some numerical experiments contribute to illustrate these concepts.
Language: English
ISSN: 2352-1465MIAR
Appears in Collections:Conference lectures

Files in This Item:
File Description SizeFormat 
  Restricted Access
516,9 kBAdobe PDFView/Open Request a copy