29.06.2013 Views

NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...

NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...

NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

A Computer-Vision Based Interface for the Control of Parameterised Music<br />

Effects<br />

Abstract<br />

It is increasingly common for musicians to<br />

route the signal from their instrument through a MIDI<br />

interface to be processed digitally by commerciallyavailable<br />

software rather than by traditional analog<br />

effects units and amplifiers. This allows for greater<br />

sonic possibilities and control.<br />

This project investigates the potential of<br />

utilizing computer vision techniques in the task of<br />

controlling this processing, i.e. parameterised music<br />

effects. The proposed system would create a flexible<br />

and intuitive interface, enabling the control of any<br />

processing being applied to an instruments signal in<br />

real-time by using head movements as an interface to<br />

generate MIDI control signals.<br />

Introduction<br />

At present, the parameters of music effects are<br />

controlled either through the knobs, sliders and faders<br />

found on MIDI interfaces or through floor-based units.<br />

Current MIDI interfaces have the limitation<br />

that in order to change a parameter, one must do so by<br />

hand. This has the implication that one hand is<br />

essentially removed from the instrument being played.<br />

Floor-based units also have knobs enabling the<br />

control of parameters. More tailored units enable the<br />

control of a single parameter through a foot-pedal that<br />

moves along one-axis (i.e. rocked back and forth from<br />

heel to toe).<br />

A computer-vision approach based on head<br />

movements would enable the control of a number of<br />

effects over three axis of motion <strong>–</strong> pitch, roll, and yaw <strong>–</strong><br />

while enabling both hands to be engaged in playing the<br />

instrument at the same time.<br />

Problems<br />

Head tracking poses two distinct problems: 1)<br />

the detection of the head in an initial frame and 2)<br />

tracking the head, or features within the face, in all<br />

subsequent frames. The human face has, by its nature, a<br />

high degree of variability in its appearance. This makes<br />

face detection a difficult problem in computer vision in<br />

comparison to other problems, where the appearance of<br />

the object to be detected and the location of the camera<br />

may be known in advance.<br />

K. Stephens<br />

Information Technology, <strong>NUI</strong> <strong>Galway</strong><br />

Supervisor: Dr. S. Redfern<br />

k.stephens2@nuigalway.ie<br />

6<br />

Although a number of head-tracking systems<br />

already exist, few are examined under or designed for<br />

all the conditions that such as system would be required<br />

to operate in. Namely:<br />

Approach<br />

• Low latency, ensuring the high level of<br />

responsiveness that would be required by a<br />

musician.<br />

• Low-level ambient lighting<br />

• High levels of dynamic lighting<br />

• Complex backgrounds where motion<br />

independent of the focus head is present<br />

• Handle partial/full occlusions of the face<br />

gracefully & scaling of face as performer<br />

moves in relation to camera<br />

• Intuitive to use<br />

As part of this project, a feature-based tracking<br />

algorithm is developed. A successful system should be<br />

one that is identity-independent i.e. can automatically<br />

detect features without any prior user-focused feature<br />

training; and, one which is capable of doing so<br />

efficiently and robustly.<br />

Feature-based tracking, involves matching<br />

local interest-points between subsequent frames to<br />

update the tracking parameters. Because local feature<br />

matching does not depend on training data, it is less<br />

sensitive to illumination and object appearance. This is<br />

in contrast to appearance-based approaches, such as 2D<br />

Active Appearance Models (AAM)[1] which, although<br />

more stable, can have difficulty generalising to<br />

unseen/untrained images.<br />

The system is developed in C++ with the<br />

OpenCV library, a library developed by Intel, optimised<br />

for Intel architecture and focused on real-time image<br />

processing.<br />

References<br />

[1] T.F.Cootes, G.J. Edwards and C.J.Taylor. "Active<br />

Appearance Models", IEEE Pattern Analysis and Machine<br />

Intelligence, Vol.23, No.6, pp.681-685, 2001

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!