28.07.2013 Views

Project Proposal (PDF) - Oxford Brookes University

Project Proposal (PDF) - Oxford Brookes University

Project Proposal (PDF) - Oxford Brookes University

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

FP7-ICT-2011-9 STREP proposal<br />

18/01/12 v1 [Dynact]<br />

Work package description<br />

Work package number 4 Start date or starting event: Month 1<br />

Work package title Data collection and feature extraction<br />

Activity type 15 RTD<br />

Participant number 1 4 5<br />

Participant short name OBU SUP DYN<br />

Person-months per participant 18 18 6<br />

Objectives<br />

The goal of this WP is to develop the infrastructure for the validation of the methodological breakthroughs<br />

developed in WP1-3 on the scenarios we isolated in WP5. To this purpose it is necessary to:<br />

– define and acquire novel test-beds in multiple modalities: synchronised video sequences from multiple<br />

cameras (stereo), monocular videos, and range camera sequences; we will focus on complex activities and<br />

multiple actions taking place at different locations in the same video;<br />

– create and manage a data repository for dissemination among the partners;<br />

– design and implement the selection and extraction of salient pieces of information (features) in all the<br />

three modalities; features are later fed to parameter estimation algorithms (imprecise EM, WPs 1,2) or<br />

discriminative models (WP3) to locate and classify actions in video sequences. State-of-the-art feature<br />

selection algorithms are designed, in both 2D (e.g. dense trajectories) and 3D.<br />

Description of work<br />

Task 4.1 – Novel benchmark datasets gathering. Video sequence test-beds depicting multiple actors<br />

performing different actions and gestures at different spatio-temporal locations, with or without missing<br />

data and occlusions, under different environmental conditions, in different scenarios (human computer<br />

interaction, smart room, surveillance, identity recognition) are collected in three different modalities: stereo,<br />

monocular video, and range cameras. A data repository is created to store and efficiently share the gathered<br />

test-bed data among all participants, and eventually the public via a secure web site, or an intranet section of<br />

the project website developed for wider dissemination purposes.<br />

Task 4.2 – Discriminative feature selection and extraction algorithms are designed starting from stateof-the-art<br />

approaches, such as dense trajectories in spatio-temporal volumes. Similar automatic feature<br />

selection algorithms are specifically designed for range data (depth maps). In this case reliable silhouettes<br />

or rough 3D volumetric representations of the moving person/body-part can be extracted. Unsupervised<br />

techniques for tracking volumetric features in an embedding space are also explored. Active Appearance<br />

Models of realistic animations are built via integration of range data by genetic algorithms.<br />

Deliverables<br />

D4.1 A demonstrator including a corpus of video sequences acquired from multiple synchronised cameras,<br />

depicting multiple actors performing different actions/activities in different S/T locations in different<br />

scenarios (human computer interaction, smart room, surveillance, identity recognition) (month 12).<br />

D4.2 A demonstrator including a corpus of range data sequences, also picturing different people performing<br />

different gestures or facial expressions at the same time, mostly in indoor scenarios (virtual animations,<br />

gaming) due to inherent limitations of range cameras (month 12).<br />

D4.3 Shared data repository storing the gathered test-bed data in the three selected modalities, to be later<br />

made available to the wider academic community and the public (month 12).<br />

D4.4 A library of routines for automatic robust discriminative feature extraction from monocular videos and<br />

synchronised 2D frames (month 15).<br />

D4.5 A library of routines for automatic robust discriminative feature extraction from range data (sequences<br />

of depth maps) (month 18).<br />

15 Please indicate one activity per work package: RTD = Research and technological development; DEM =<br />

Demonstration; MGT = Management of the consortium.<br />

<strong>Proposal</strong> Part B: page [31] of [67]

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!