Learning to translate between real world and simulated 3D sensors while transferring task models
- Learning-based vision tasks are usually specialized on the sensor technology for which data has been labeled. The knowledge of a learned model is simply useless when it comes to data which differs from the data on which the model has been initially trained or if the model should be applied to a totally different imaging or sensor source. New labeled data has to be acquired on which a new model can be trained. Depending on the sensor, this can even get more complicated when the sensor data becomes more abstract and hard to be interpreted and labeled by humans. To enable reuse of models trained for a specific task across different sensors minimizes the data acquisition effort. Therefore, this work focuses on learning sensor models and translating between them, thus aiming for sensor interoperability. We show that even for the complex task of human pose estimation from 3D depth data recorded with different sensors, i.e. a simulated and a Kinect 2TM depth sensor, human pose estimation can greatly improve by translating between sensor models without modifying the original task model. This process especially benefits sensors and applications for which labels and models are difficult if at all possible to retrieve from raw sensor data.
Author of HS Reutlingen | Essich, Michael; Ludl, Dennis; Gulde, Thomas; Curio, Cristóbal |
---|---|
DOI: | https://doi.org/10.1109/3DV.2019.00080 |
ISBN: | 978-1-72813-131-3 |
Erschienen in: | 2019 International Conference on 3D Vision : 3DV 2019 : Quebec, Canada, 15-18 September 2019 : proceedings |
Publisher: | IEEE |
Place of publication: | Piscataway, NJ |
Editor: | Denis Laurendeau |
Document Type: | Conference proceeding |
Language: | English |
Publication year: | 2019 |
Page Number: | 9 |
First Page: | 681 |
Last Page: | 689 |
DDC classes: | 004 Informatik |
Open access?: | Nein |
Licence (German): | In Copyright - Urheberrechtlich geschützt |