Reduced egomotion estimation drift using omnidirectional views
| dc.contributor.author | Baştanlar, Yalın | |
| dc.coverage.doi | 10.5565/rev/elcvia.564 | |
| dc.date.accessioned | 2017-06-12T13:15:31Z | |
| dc.date.available | 2017-06-12T13:15:31Z | |
| dc.date.issued | 2014 | |
| dc.description.abstract | Estimation of camera motion from a given image sequence is a common task for multi-view 3D computer vision applications. Salient features (lines, corners etc.) in the images are used to estimate the motion of the camera, also called egomotion. This estimation suffers from an error built-up as the length of the image sequence increases and this causes a drift in the estimated position. In this letter, this phenomenon is demonstrated and an approach to improve the estimation accuracy is proposed. The main idea of the proposed method is using an omnidirectional camera (360° horizontal field of view) in addition to a conventional (perspective) camera. Taking advantage of the correspondences between the omnidirectional and perspective images, the accuracy of camera position estimates can be improved. In our work, we adopt the sequential structure-from-motion approach which starts with estimating the motion between first two views and more views are added one by one. We automatically match points between omnidirectional and perspective views. Point correspondences are used for the estimation of epipolar geometry, followed by the reconstruction of 3D points with iterative linear triangulation. In addition, we calibrate our cameras using sphere camera model which covers both omnidirectional and perspective cameras. This enables us to treat the cameras in the same way at any step of structure-from-motion. We performed simulated and real image experiments to compare the estimation accuracy when only perspective views are used and when an omnidirectional view is added. Results show that the proposed idea of adding omnidirectional views reduces the drift in egomotion estimation. | en_US |
| dc.identifier.citation | Bastanlar, Y. (2014). Reduced egomotion estimation drift using omnidirectional views. Electronic Letters on Computer Vision and Image Analysis, 13(3), 1-12. doi:10.5565/rev/elcvia.564 | en_US |
| dc.identifier.doi | 10.5565/rev/elcvia.564 | en_US |
| dc.identifier.doi | 10.5565/rev/elcvia.564 | |
| dc.identifier.issn | 1577-5097 | |
| dc.identifier.scopus | 2-s2.0-84919771361 | |
| dc.identifier.uri | https://doi.org/10.5565/rev/elcvia.564 | |
| dc.identifier.uri | https://hdl.handle.net/11147/5744 | |
| dc.language.iso | en | en_US |
| dc.publisher | Centre de Visio per Computador | en_US |
| dc.relation.ispartof | Electronic Letters on Computer Vision and Image Analysis | en_US |
| dc.rights | info:eu-repo/semantics/openAccess | en_US |
| dc.subject | Egomotion estimation | en_US |
| dc.subject | Omnidirectional cameras | en_US |
| dc.subject | Structure-from-motion | en_US |
| dc.subject | Visual odometry | en_US |
| dc.title | Reduced egomotion estimation drift using omnidirectional views | en_US |
| dc.type | Article | en_US |
| dspace.entity.type | Publication | |
| gdc.author.institutional | Baştanlar, Yalın | |
| gdc.author.yokid | 176747 | |
| gdc.bip.impulseclass | C5 | |
| gdc.bip.influenceclass | C5 | |
| gdc.bip.popularityclass | C5 | |
| gdc.coar.access | open access | |
| gdc.coar.type | text::journal::journal article | |
| gdc.collaboration.industrial | false | |
| gdc.description.department | İzmir Institute of Technology. Computer Engineering | en_US |
| gdc.description.endpage | 12 | en_US |
| gdc.description.issue | 3 | en_US |
| gdc.description.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | Q3 | |
| gdc.description.startpage | 1 | en_US |
| gdc.description.volume | 13 | en_US |
| gdc.description.wosquality | N/A | |
| gdc.identifier.openalex | W1989231208 | |
| gdc.index.type | Scopus | |
| gdc.oaire.accesstype | GOLD | |
| gdc.oaire.diamondjournal | false | |
| gdc.oaire.impulse | 0.0 | |
| gdc.oaire.influence | 2.635068E-9 | |
| gdc.oaire.isgreen | true | |
| gdc.oaire.keywords | FOS: Computer and information sciences | |
| gdc.oaire.keywords | Visual Odometry | |
| gdc.oaire.keywords | Computer engineering. Computer hardware | |
| gdc.oaire.keywords | Omnidirectional Cameras, Structure-from-motion, Egomotion Estimation, Visual Odometry | |
| gdc.oaire.keywords | Computer Vision | |
| gdc.oaire.keywords | Computer Vision and Pattern Recognition (cs.CV) | |
| gdc.oaire.keywords | Visual odometry | |
| gdc.oaire.keywords | Egomotion estimation | |
| gdc.oaire.keywords | Computer Science - Computer Vision and Pattern Recognition | |
| gdc.oaire.keywords | QA75.5-76.95 | |
| gdc.oaire.keywords | TK7885-7895 | |
| gdc.oaire.keywords | Computer Science - Robotics | |
| gdc.oaire.keywords | Omnidirectional Cameras | |
| gdc.oaire.keywords | Electronic computers. Computer science | |
| gdc.oaire.keywords | Egomotion Estimation | |
| gdc.oaire.keywords | Omnidirectional cameras | |
| gdc.oaire.keywords | Robotics (cs.RO) | |
| gdc.oaire.keywords | Structure-from-motion | |
| gdc.oaire.popularity | 7.742206E-10 | |
| gdc.oaire.publicfunded | false | |
| gdc.oaire.sciencefields | 02 engineering and technology | |
| gdc.oaire.sciencefields | 0202 electrical engineering, electronic engineering, information engineering | |
| gdc.openalex.collaboration | National | |
| gdc.openalex.fwci | 0.24114154 | |
| gdc.openalex.normalizedpercentile | 0.57 | |
| gdc.opencitations.count | 0 | |
| gdc.plumx.mendeley | 2 | |
| gdc.plumx.scopuscites | 0 | |
| gdc.scopus.citedcount | 0 | |
| relation.isAuthorOfPublication.latestForDiscovery | 7f75e80a-0468-490d-ba2e-498de80b7217 | |
| relation.isOrgUnitOfPublication.latestForDiscovery | 9af2b05f-28ac-4014-8abe-a4dfe192da5e |
