i3DPost Multi-view Human Action Datasets |
This is a webpage to distribute multi-view/3D human action/interaction datasets created in a cooperation between University of Surrey and CERTH-ITI within the i3DPost project. All details about the datasets are described in this paper: N. Gkalelis, H. Kim, A. Hilton, N. Nikolaidis and I. Pitas, "The i3DPost multi-view and 3D human action/interaction," Proc. CVMP, pp. 159-168, 2009 PDF In order to access the full datasets, please read this license agreement and send email Dr. Hansung Kim with the following information if you agree: Your name/affiliation, name/email of your supervisor (if you are a student) Access to the datasets (ID/PW required) |
If you get a permission to access, the following datasets are provided. - Multi-view datasets - Synchronised/uncompressed-HD 8 view image sequences of 13 actions of 8 people (104 total) - Sample videos Actors/Actresses: Man1, Woman1, Man2, Man3, Woman2, Man4, Man5, Man6, Actions: Walk, Run, Jump, Bend, Hand-wave, Jump-in-place, Sit-StandUp, Run-fall, Walk-sit, Run-jump-walk, Handshake, Pull, Facial-expressions - Background images for matting - Camera calibration parameters (intrinsic and extrinsic) for 3D reconstruction - 3D mesh models - All frames and actions of above datasets in ntri ascii format The models were reconstructed using a global optimisation method: J. Starck and A. Hilton, “Surface capture for performance based animation,” IEEE Comp. Graph. Appl., 27(3), pp.21–31, 2007. PDF |
Centre for Vision, Speech and Signal Processing, University Of Surrey, Guildford, UK Artificial Intelligence & Information Analysis Lab, CERTH-ITI, Thessaloniki, GREECE i3DPost, EU Project under Framework 7 ICT Programme To contact us: Dr. Hansung Kim (h.kim@surrey.ac.uk) |