Multiple UAV routing for re-planning under dynamic environment using deep reinforcement learning동적 임무 환경 변화를 고려한 심층 강화학습 기반 다수 무인기 경로 재계획 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 148
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorAhn, Jaemyung-
dc.contributor.advisor안재명-
dc.contributor.authorLee, Dong-Ho-
dc.date.accessioned2023-06-26T19:32:23Z-
dc.date.available2023-06-26T19:32:23Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1008447&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/309698-
dc.description학위논문(석사) - 한국과학기술원 : 항공우주공학과, 2022.8,[iii, 46 p. :]-
dc.description.abstractThis thesis studies an autonomous mission planning algorithm that enables unmanned aerial vehicles (UAVs) to autonomously re-plan missions without human direct intervention under various dynamic mission environment changes, such as the creation of new missions, cancellation or change of existing missions. In this study, the mission re-planning problem is developed based on the Team Orienteering Problem (TOP), one of the variants of the Vehicle Routing Problem with Profits (VRPP). Existing deep reinforcement learning (DRL) based methods have largely focused on learning heuristics for Vehicle Routing Problem (VRP) and its variants that intrinsically have vehicles departing from a given depot and returning to that depot. While this setting is necessary to plan routing missions in advance, it needs to be expanded to cope with mission re-planning scenarios where vehicles are located away from the depot at the start. Additionally, many real-life re-planning situations are subject to a fuel constraint on each vehicle, which is likely to have variable remaining fuel. Therefore, this thesis investigates a Multiple-Start TOP (MSTOP), in which vehicles begin at multiple random locations, travel to maximize the total prizes, and arrive at the given depot, while satisfying fuel constraints. To solve MSTOP, this thesis proposes a methodology consisting of self-attention mechanism on each partial tour, and encoder-decoder attention mechanism between partial tour and remaining nodes. The proposed DRL-based method produces a suboptimal solution comparable to the existing meta-heuristic techniques, even for more complex problems. Furthermore, several case studies are presented to demonstrate the performance of the proposed model and solution procedure.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectUAV autonomous mission planning▼aTeam orienteering problem (TOP)▼aMixed integer linear programming (MILP)▼aDeep reinforcement learning (DRL)-
dc.subject무인기 자율 임무 계획▼a차량 경로 결정 문제▼a혼합정수계획법▼a심층 강화학습-
dc.titleMultiple UAV routing for re-planning under dynamic environment using deep reinforcement learning-
dc.title.alternative동적 임무 환경 변화를 고려한 심층 강화학습 기반 다수 무인기 경로 재계획 연구-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :항공우주공학과,-
dc.contributor.alternativeauthor이동호-
Appears in Collection
AE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0