SpherePHD: Applying CNNs on a Spherical PolyHeDron Representation of 360° images

Cited 72 time in webofscience Cited 37 time in scopus
  • Hit : 333
  • Download : 0
Omni-directional cameras have many advantages over conventional cameras in that they have a much wider field-of-view (FOV). Accordingly, several approaches have been proposed recently to apply convolutional neural networks (CNNs) to omni-directional images for various visual tasks. However, most of them use image representations defined in the Euclidean space after transforming the omni-directional views originally formed in the non-Euclidean space. This transformation leads to shape distortion due to nonuniform spatial resolving power and the loss of continuity. These effects make existing convolution kernels experience difficulties in extracting meaningful information. This paper presents a novel method to resolve such problems of applying CNNs to omni-directional images. The proposed method utilizes a spherical polyhedron to represent omni-directional views. This method minimizes the variance of the spatial resolving power on the sphere surface, and includes new convolution and pooling methods for the proposed representation. The proposed method can also be adopted by any existing CNN-based methods. The feasibility of the proposed method is demonstrated through classification, detection, and semantic segmentation tasks with synthetic and real datasets.
Publisher
IEEE CVPR
Issue Date
2019-06-16
Language
English
Citation

IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, pp.9173 - 9181

DOI
10.1109/CVPR.2019.00940
URI
http://hdl.handle.net/10203/263397
Appears in Collection
ME-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 72 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0