Deep learning-based sequential image contrast enhancement for enhancing visual odometry영상 기반 위치 추정 알고리즘의 향상을 위한 딥 러닝 기반 순차 이미지 대비 향상

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 2
  • Download : 0
This paper proposes a novel image processing network designed to enhance image contrast for improved performance in visual odometry. Pose estimation results in visual odometry are often susceptible to challenges arising from image illumination condition and the quantity of extracted features. To address these issues, the proposed network integrates dilated convolution layers and Convolutional Graphed Recurrent Units (GRU), leveraging spatio-temporal information within the image sequences. Significantly, the loss function is delicately designed to enhance image contrast while preserving both structural and sequential informational content during training. Leveraging SSIM loss, contrast loss, and optical flow loss, the training process aims to enhance image contrast to amplify feature points while preserving innate image details, contributing to improved visual odometry results. The network’s training is conducted by unsupervised learning with a custom image dataset collected from alleys and parking lots inside the N1 building at KAIST. The proposed network has three principal contributions: robust tracking, enhanced accuracy, and real-time performance. Validation of the network’s performance contains a comparison of pose estimation results between raw and processed images using open-source visual SLAM. Robustness is validated through comparisons between raw and processed images, using sets of images acquired in tunnel environments. Tracking performance of visual slam has improved, demonstarting the ability of proposed network. The ETH3D dataset, a well-established benchmark, is also employed to validate the network’s enhanced tracking capabilities. Furthermore, the experiments utilizing KITTI dataset highlight an enhancement in accuracy of visual odometry, proving the network’s proficiency in enhancing the contrast of image. Lastly, a UAV flight experiment is conducted to validate the real-time applicability of the proposed network. Navigation, collision avoidance, and control modules are implemented in the embedded computer, proving real-time pose estimation results. This comprehensive examination demonstrates the ability of the proposed image processing network for enhancing visual odometry.
Advisors
심현철researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[v, 38 p. :]

Keywords

영상 기반 위치 추정▼a딥 러닝▼a대비 향상▼a팽창 합성곱▼a합성곱 게이트 순환 유; Visual odometry▼aDeep learning▼aContrast enhancement▼aDilated convolution▼aConvolutional gated recurrent unit

URI
http://hdl.handle.net/10203/321623
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1097195&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0