Empreu aquest identificador per citar o enllaçar aquest ítem:
http://hdl.handle.net/10609/148336
Títol: | Optimized path dataset representation |
Autoria: | Pérez Cervera, Daniel |
Tutor: | Perez-Roses, Hebert |
Resum: | Path-based data is increasingly used in mobility applications. This type of data has multiple uses, among them studying and predicting travelers behavior. Paths are usually represented as a sequence of vertices on a road network graph, sequences of geographical coordinates and omitting its temporal component. In this work, we propose a novel, generic and optimized representation of Path Datasets which can be used for any type of data. Such approach is designed to be efficient for collecting, storing and processing path-based data associated to Road Network trips. This approach is novel, as it considers representing paths collectively by origin vertex using tries to index paths uniquely. In particular, for path storage, where we use two effective techniques: DFUDS and Adaptive Edge Offset Compression. We evaluate this approach with other baseline approaches using synthetically generated Path Datasets, to assess the gains in performance (computation time, memory, and storage size) and explain them mainly through a quantity we define, the overlap θ. Results show this approach is more efficient for higher overlap among the paths stored, and never worse than storing for the generated path datasets. The only exception is for the collection stage, in which time and space complexity are higher in some situations. |
Paraules clau: | road network path dataset trie |
Tipus de document: | info:eu-repo/semantics/masterThesis |
Data de publicació: | 2-jul-2023 |
Llicència de publicació: | http://creativecommons.org/licenses/by-nc-nd/3.0/es/ |
Apareix a les col·leccions: | Treballs finals de carrera, treballs de recerca, etc. |
Arxius per aquest ítem:
Arxiu | Descripció | Mida | Format | |
---|---|---|---|---|
dperezceFMDP0723report.pdf | Report of TFM | 2,9 MB | Adobe PDF | Veure/Obrir |
Comparteix:
Aquest ítem està subjecte a una llicència de Creative CommonsLlicència Creative Commons