NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...
NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...
NUI Galway – UL Alliance First Annual ENGINEERING AND - ARAN ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
A Computer-Vision Based Interface for the Control of Parameterised Music<br />
Effects<br />
Abstract<br />
It is increasingly common for musicians to<br />
route the signal from their instrument through a MIDI<br />
interface to be processed digitally by commerciallyavailable<br />
software rather than by traditional analog<br />
effects units and amplifiers. This allows for greater<br />
sonic possibilities and control.<br />
This project investigates the potential of<br />
utilizing computer vision techniques in the task of<br />
controlling this processing, i.e. parameterised music<br />
effects. The proposed system would create a flexible<br />
and intuitive interface, enabling the control of any<br />
processing being applied to an instruments signal in<br />
real-time by using head movements as an interface to<br />
generate MIDI control signals.<br />
Introduction<br />
At present, the parameters of music effects are<br />
controlled either through the knobs, sliders and faders<br />
found on MIDI interfaces or through floor-based units.<br />
Current MIDI interfaces have the limitation<br />
that in order to change a parameter, one must do so by<br />
hand. This has the implication that one hand is<br />
essentially removed from the instrument being played.<br />
Floor-based units also have knobs enabling the<br />
control of parameters. More tailored units enable the<br />
control of a single parameter through a foot-pedal that<br />
moves along one-axis (i.e. rocked back and forth from<br />
heel to toe).<br />
A computer-vision approach based on head<br />
movements would enable the control of a number of<br />
effects over three axis of motion <strong>–</strong> pitch, roll, and yaw <strong>–</strong><br />
while enabling both hands to be engaged in playing the<br />
instrument at the same time.<br />
Problems<br />
Head tracking poses two distinct problems: 1)<br />
the detection of the head in an initial frame and 2)<br />
tracking the head, or features within the face, in all<br />
subsequent frames. The human face has, by its nature, a<br />
high degree of variability in its appearance. This makes<br />
face detection a difficult problem in computer vision in<br />
comparison to other problems, where the appearance of<br />
the object to be detected and the location of the camera<br />
may be known in advance.<br />
K. Stephens<br />
Information Technology, <strong>NUI</strong> <strong>Galway</strong><br />
Supervisor: Dr. S. Redfern<br />
k.stephens2@nuigalway.ie<br />
6<br />
Although a number of head-tracking systems<br />
already exist, few are examined under or designed for<br />
all the conditions that such as system would be required<br />
to operate in. Namely:<br />
Approach<br />
• Low latency, ensuring the high level of<br />
responsiveness that would be required by a<br />
musician.<br />
• Low-level ambient lighting<br />
• High levels of dynamic lighting<br />
• Complex backgrounds where motion<br />
independent of the focus head is present<br />
• Handle partial/full occlusions of the face<br />
gracefully & scaling of face as performer<br />
moves in relation to camera<br />
• Intuitive to use<br />
As part of this project, a feature-based tracking<br />
algorithm is developed. A successful system should be<br />
one that is identity-independent i.e. can automatically<br />
detect features without any prior user-focused feature<br />
training; and, one which is capable of doing so<br />
efficiently and robustly.<br />
Feature-based tracking, involves matching<br />
local interest-points between subsequent frames to<br />
update the tracking parameters. Because local feature<br />
matching does not depend on training data, it is less<br />
sensitive to illumination and object appearance. This is<br />
in contrast to appearance-based approaches, such as 2D<br />
Active Appearance Models (AAM)[1] which, although<br />
more stable, can have difficulty generalising to<br />
unseen/untrained images.<br />
The system is developed in C++ with the<br />
OpenCV library, a library developed by Intel, optimised<br />
for Intel architecture and focused on real-time image<br />
processing.<br />
References<br />
[1] T.F.Cootes, G.J. Edwards and C.J.Taylor. "Active<br />
Appearance Models", IEEE Pattern Analysis and Machine<br />
Intelligence, Vol.23, No.6, pp.681-685, 2001