PopUp: Reconstructing 3D video using particle filtering to aggregate crowd responses

Cited 4 time in webofscience Cited 7 time in scopus
  • Hit : 225
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorSong, Jean Yko
dc.contributor.authorLemmer, Stephan Jko
dc.contributor.authorLiu, Michael Xieyangko
dc.contributor.authorYan, Shiyanko
dc.contributor.authorKim, Juhoko
dc.contributor.authorCorso, Jason Jko
dc.contributor.authorLasecki, Walter Sko
dc.date.accessioned2019-12-13T11:28:16Z-
dc.date.available2019-12-13T11:28:16Z-
dc.date.created2019-12-02-
dc.date.created2019-12-02-
dc.date.issued2019-03-18-
dc.identifier.citation24th ACM International Conference on Intelligent User Interfaces, IUI 2019, pp.558 - 569-
dc.identifier.urihttp://hdl.handle.net/10203/269512-
dc.description.abstractCollecting a sufficient amount of 3D training data for autonomous vehicles to handle rare, but critical, traffic events (e.g., collisions) may take decades of deployment. Abundant video data of such events from municipal traffic cameras and video sharing sites (e.g., YouTube) could provide a potential alternative, but generating realistic training data in the form of 3D video reconstructions is a challenging task beyond the current capabilities of computer vision. Crowdsourcing the annotation of necessary information could bridge this gap, but the level of accuracy required to obtain usable reconstructions makes this task nearly impossible for non-experts. In this paper, we propose a novel hybrid intelligence method that combines annotations from workers viewing different instances (video frames) of the same target (3D object), and uses particle filtering to aggregate responses. Our approach can leveraging temporal dependencies between video frames, enabling higher quality through more aggressive filtering. The proposed method results in a 33% reduction in the relative error of position estimation compared to a state-of-the-art baseline. Moreover, our method enables skipping (self-filtering) challenging annotations, reducing the total annotation time for hard-to-annotate frames by 16%. Our approach provides a generalizable means of aggregating more accurate crowd responses in settings where annotation is especially challenging or error-prone.-
dc.languageEnglish-
dc.publisherAssociation for Computing Machinery-
dc.titlePopUp: Reconstructing 3D video using particle filtering to aggregate crowd responses-
dc.typeConference-
dc.identifier.wosid000481595500060-
dc.identifier.scopusid2-s2.0-85065564793-
dc.type.rimsCONF-
dc.citation.beginningpage558-
dc.citation.endingpage569-
dc.citation.publicationname24th ACM International Conference on Intelligent User Interfaces, IUI 2019-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationMarina del Ray, California-
dc.identifier.doi10.1145/3301275.3302305-
dc.contributor.localauthorKim, Juho-
dc.contributor.nonIdAuthorSong, Jean Y-
dc.contributor.nonIdAuthorLemmer, Stephan J-
dc.contributor.nonIdAuthorLiu, Michael Xieyang-
dc.contributor.nonIdAuthorYan, Shiyan-
dc.contributor.nonIdAuthorCorso, Jason J-
dc.contributor.nonIdAuthorLasecki, Walter S-
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 4 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0