Imitation Learning-Based Unmanned Aerial Vehicle Planning for Multitarget Reconnaissance Under Uncertainty

Cited 7 time in webofscience Cited 9 time in scopus
  • Hit : 569
  • Download : 0
This paper proposes a planning framework for a multitarget unmanned aerial vehicle (UAV) reconnaissance mission under target position uncertainty based on the imitation learning of a policy network. A problem in planning the flight path and task sequence for the reconnaissance of given targets is formulated as a Markov decision process (MDP). A deep neural classifier is introduced as a policy network to predict the probability over admissible immediate actions at each mission state. A set of MDPs with specified target locations (training set) are constructed and solved offline to generate the segments of underlying optimal policy, which are used to train the policy network. The trained policy networks are used for online (runtime) planning of the UAV conducting reconnaissance for arbitrary target locations by providing the near-optimal action corresponding to specified target locations and the current state. A numerical case study has been conducted to demonstrate the effectiveness of the proposed approach from the perspective of its classification performance and the expected return in comparison with a baseline online planning algorithm.
Publisher
AMER INST AERONAUTICS ASTRONAUTICS
Issue Date
2020-01
Language
English
Article Type
Article
Citation

JOURNAL OF AEROSPACE INFORMATION SYSTEMS, v.17, no.1, pp.36 - 50

ISSN
1940-3151
DOI
10.2514/1.I010759
URI
http://hdl.handle.net/10203/272207
Appears in Collection
AE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 7 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0