Texte intégral / Full text (pdf, 20 MiB) - Infoscience - EPFL
Texte intégral / Full text (pdf, 20 MiB) - Infoscience - EPFL
Texte intégral / Full text (pdf, 20 MiB) - Infoscience - EPFL
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
CHAPTER 3<br />
VRET For Social Phobia<br />
Much work has already been conducted with regards to the use of VR as therapeutic tool<br />
for the treatment of phobias. Many types of phobias have thus been tackled such as acrophobia,<br />
arachnophobia, and claustrophobia. It is quite simple to simulate heights, spiders, or<br />
closed spaces in VEs. On the other hand, it is much harder to simulate humans. Indeed, as<br />
human beings, we are experts in human being representation and movement. For this reason,<br />
much less work has been undergone in social phobia. VRET for social phobia, and more<br />
specifically, for fear of public speaking is only about a decade old.<br />
In this chapter, we first present the various scenarios we have created, including environments,<br />
characters and scenario scripting. We then present a method we have developed in<br />
order to obtain interaction between the user and a character using eye-tracking data. We then<br />
discuss how this has motivated us to further work on character gaze behaviors.<br />
3.1 Preliminaries<br />
The first part of this thesis is based on previous work which has been undergone in the Virtual<br />
Reality Laboratory (VRLab, <strong>EPFL</strong>). More specifically, the platform which we use has been<br />
developed by Herbelin during his PhD thesis [Herbelin, <strong>20</strong>05].<br />
This framework allows for the real-time animation of a small group of characters. They<br />
are endowed with gaze control and facial animation. The platform works with Python scripts.<br />
These scripts allow interactive control of characters in order to make them talk for example;<br />
this actually consists of playing a pre-recorded sentence and animating the character’s face<br />
and eyes accordingly.<br />
35