28.01.2015 Views

s Voice to Make Human-Robot Interaction More Engaging

s Voice to Make Human-Robot Interaction More Engaging

s Voice to Make Human-Robot Interaction More Engaging

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Using a <strong>Robot</strong>'s <strong>Voice</strong> <strong>to</strong> <strong>Make</strong> <strong>Human</strong>-<strong>Robot</strong> <strong>Interaction</strong><br />

<strong>More</strong> <strong>Engaging</strong><br />

Hans van de Kamp<br />

University of Twente<br />

P.O. Box 217, 7500AE Enschede<br />

The Netherlands<br />

h.vandekamp@student.utwente.nl<br />

ABSTRACT<br />

Nowadays a robot is becoming more than just a machine, the<br />

robot becomes an interaction partner. A human needs <strong>to</strong> be<br />

engaged <strong>to</strong> interact with the robot. This paper is about an<br />

experiment on robot voices in a task-based environment. The<br />

goal was <strong>to</strong> determine the influence of the robot's voice on the<br />

way humans are engaged or interested <strong>to</strong> perform a certain<br />

task. This research is contributing <strong>to</strong> the <strong>to</strong>pic of engagement<br />

in human-robot interaction with different voice styles. The<br />

participant is asked <strong>to</strong> perform six small assignments <strong>to</strong><br />

measure the effects of the different voices; a human-like voice<br />

(N=10) and a machine-like or mechanical voice (N=11).<br />

There were some significant differences between the two<br />

voices, mostly related <strong>to</strong> the likeability of the robot. The<br />

differences between the voices in terms of interest or<br />

engagement turned out <strong>to</strong> be minimal and not significant.<br />

Keywords<br />

<strong>Robot</strong> voice, challenging, engagement, human-robot<br />

interaction, task interest, robot interest<br />

1. INTRODUCTION<br />

Different frameworks for human-robot interaction have been<br />

created in the past years [1]. Most of them are trying <strong>to</strong><br />

improve human-robot interaction by incorporating human<br />

behavior and human personality traits in robots [7, 16]. <strong>Voice</strong><br />

is an important fac<strong>to</strong>r in human personality [8], therefore<br />

many robots use a human-like voice <strong>to</strong> interact with humans.<br />

Creating a human-like interaction partner has proven <strong>to</strong> be<br />

valuable in human-robot interaction [4, 10] in terms of<br />

effectiveness and efficiency [11, 15]. This research will not<br />

evaluate human-robot interaction by measuring task<br />

effectiveness or task efficiency, but it will evaluate humanrobot<br />

interaction by the way humans are engaged [12] while<br />

performing a certain assignment. This research aims <strong>to</strong><br />

provide insight on the relationship between a robot’s voice<br />

and engagement [12] in terms of interest.<br />

Permission <strong>to</strong> make digital or hard copies of all or part of this work for<br />

personal or classroom use is granted without fee provided that copies are<br />

not made or distributed for profit or commercial advantage and that<br />

copies bear this notice and the full citation on the first page. To copy<br />

otherwise, or republish, <strong>to</strong> post on servers or <strong>to</strong> redistribute <strong>to</strong> lists,<br />

requires prior specific permission and/or a fee.<br />

18 th Twente Student Conference on IT, January 25 th , 2013, Enschede, The<br />

Netherlands.<br />

Copyright 2013, University of Twente, Faculty of Electrical Engineering,<br />

Mathematics and Computer Science.<br />

1.1 Problem statement<br />

Other research [4] focuses on models in which robots mirror<br />

the participants’ behavior <strong>to</strong> create a more human-like<br />

interaction partner. However, most of that research does not<br />

focus on engaging a human. This paper focuses only on<br />

engaging a participant in human-robot interaction. <strong>Engaging</strong><br />

is related <strong>to</strong> the concept of interest as it describes attentional<br />

and emotional involvement [12]. The goal of this research is<br />

<strong>to</strong> determine the influence of a robot’s voice on the level of<br />

engagement in terms of interest [12] in human-robot<br />

interaction.<br />

It seems normal <strong>to</strong> use a more human-like robot <strong>to</strong> improve<br />

human-robot interaction, assuming that the robot is perceived<br />

as human-like. However, it is a fact that most robots are far<br />

from being like a human. If a human perceives the robot as<br />

artificial and not as human-like, it might influence the<br />

expectancy of the robot voice.<br />

If a robot is perceived as artificial and uses a mechanical voice<br />

instead of the human-like voice, does this influence the degree<br />

of which humans are engaged <strong>to</strong> perform a certain task Are<br />

humans more interested in a robot with a mechanical voice<br />

instead of a human-like voice<br />

Both questions are important <strong>to</strong> understand the relationship<br />

between humans and robots in human-robot interaction. Or in<br />

this paper more specific: the relationship between voice and<br />

engagement. With the previous questions in mind, the<br />

following research questions can be created.<br />

1.2 Research Questions<br />

To investigate the relationship between a robot’s voice and<br />

participant interest <strong>to</strong>wards the robot or task a few questions<br />

need <strong>to</strong> be answered. The main research question is:<br />

Does a robot with a humanlike voice make human-robot<br />

interaction more engaging than a mechanical voice<br />

Other research questions will be used <strong>to</strong> find an answer <strong>to</strong> the<br />

main question:<br />

- How can we find out if a person is engaged or<br />

interested in the robot or task while interacting with<br />

a robot<br />

- Does a robot’s voice influence the way participants<br />

are engaged <strong>to</strong> perform a task in terms of interest<br />

Which can be divided in<strong>to</strong> two more questions:<br />

o<br />

o<br />

Does a robot’s voice influence the way humans<br />

are interested in the robot<br />

Does a robot’s voice influence the way humans<br />

are interested in the task


2. RELATED WORK<br />

Research on human-robot interaction is often about creating a<br />

lifelike interaction partner. Therefore a lot of research is<br />

conducted with a human-like voice only, of which some are<br />

mentioned in the introduction. Important related work <strong>to</strong> this<br />

research is conducted by Walters et al. Rich and Sidner also<br />

wrote about the concept of engagement in human-robot<br />

interaction. However, they did not use different voice styles or<br />

robot appearances like Walters et al.<br />

A paper on robot appearance and personality by Walters et al.<br />

[17] investigated people’s perceptions of different robot<br />

appearances. The research uses the definitions of robot<br />

appearance based on the definitions of Gong and Nass [5] and<br />

MacDorman and Ishiguro [9]. In this experiment the robot<br />

appearances are referred <strong>to</strong> as human-like or machine-like.<br />

This paper describes the relation between interest and robot<br />

voices, it does not talk about appearance preferences in<br />

general.<br />

years with an average of 21,6 years. The second oldest<br />

participant was 24 years old.<br />

The experiment using the human voice had 10 participants<br />

and the experiment using the robot voice had 11. 20 of those<br />

participants were students of exact sciences such as Computer<br />

Science, Electrical Engineering, Mathematics or a master in<br />

the same area. The other student studied Climate &<br />

management and was apparently lost.. Because of the large<br />

amount of students with a technical background only 4<br />

participants were female. They were equally divided over<br />

both conditions.<br />

Other research by Walters et al. [18] used different robot<br />

voices similar <strong>to</strong> the experiment described in this paper.<br />

There has also been research on the gender of the robot.<br />

Siegel et al. [13] conducted research <strong>to</strong> determine the<br />

preferences of males and females on robot gender. The<br />

research showed that females generally did not have a<br />

preference of robot gender. However, males seem <strong>to</strong> have a<br />

preference of a female robot gender.<br />

Research performed by Crowell et al. [3] indicate that the<br />

perceived gender of the robot may influence human sexrelated<br />

characteristics.<br />

3. METHODOLOGY<br />

In this experiment the Magabot robot is used. A pho<strong>to</strong> of the<br />

robot is shown in the figure at the right. The robot is small<br />

(less than 1 meter tall) and offers a platform on which a lap<strong>to</strong>p<br />

can be placed.<br />

The robot will be controlled using a Wizard of Oz technique<br />

because of communication problems between the robot and<br />

Flash. The robot will drive on a platform between two tables,<br />

which will be explained in section 3.3. The voice and eyes are<br />

programmed in Flash (ActionScript). The robot used a<br />

predefined script in the form of a Flash timeline. The timeline<br />

was divided in<strong>to</strong> several segments <strong>to</strong> allow a researcher <strong>to</strong><br />

control the robot's script.<br />

As seen in the figure at the right, the lap<strong>to</strong>p displayed the<br />

robot's eyes. The robot eyes and effects such as blinking were<br />

used <strong>to</strong> make the robot more life-like. In the experiment the<br />

robot eyes were used <strong>to</strong> look at objects.<br />

Based on the research conducted by Siegel et al. [13], the<br />

robot had a female voice. Both the human-like and machinelike<br />

voice were female (or female-like) voices. The used<br />

voices are discussed in section 3.2. The robot itself had no<br />

gender. The gender was determined by the voice of the robot.<br />

The robot introduced herself as Jane.<br />

3.1 Participants<br />

A <strong>to</strong>tal of 21 participants <strong>to</strong>ok part in the experiment using a<br />

between subjects design. Their age varied between 19 and 31<br />

Figure 1: the Magabot with lap<strong>to</strong>p and eyes used in the<br />

experiment.<br />

3.2 <strong>Voice</strong><br />

Two different female voices were used in the experiment. The<br />

robot used a predefined script (in English) <strong>to</strong> communicate<br />

with the participants. Both voices were created with the<br />

MARY Text-<strong>to</strong>-Speech system. The first group of participants<br />

(N=10) interacted with a robot using a synthesized human-like<br />

female voice. The second group of participants (N=11)<br />

interacted with a robot using a synthesized machine-like or<br />

mechanical-like female voice. The used voice was in MARY<br />

called 'cmu-slt-hsmm'. The second group had a robot filter <strong>to</strong><br />

create a more machine-like or mechanical voice.<br />

3.3 Experiment Setup<br />

The conducted experiment was combined with research about<br />

proxemics in human-robot interaction, or more specific the<br />

relationship between a robot's voice and proxemics. The<br />

experiment was divided in three parts.<br />

In the first part the robot tells its name and asks the participant<br />

<strong>to</strong> come closer. This part was needed for the other experiment<br />

and is also a good introduction of the robot. The participant<br />

walks <strong>to</strong>wards the robot and the robot then tells the participant<br />

<strong>to</strong> do some simple tasks. The participant is asked <strong>to</strong> take a seat


at the table. The robot drives <strong>to</strong> the other side of the table,<br />

making sure the participant is sitting opposite of the robot.<br />

For the second part of the experiment, the participant<br />

performs six simple assignments. On the table are six cards<br />

with letters forming the word ‘thanks’ and also six (empty)<br />

numbered boxes. Three of the cards are faced up and show the<br />

letter, the others show the backside which is colored (red,<br />

green and blue). The six assignments are small assignments<br />

such as ‘move the letter N <strong>to</strong> box 2’, ‘swap the letter N with<br />

the letter H’ or ‘please turn over all colored squares’. The start<br />

of each of the assignments was triggered by hand using a<br />

Wizard of Oz technique. After some assignments the<br />

participant tells the robot how many boxes are left empty <strong>to</strong><br />

enforce interaction with the robot. After moving all cards the<br />

robot asks the participant <strong>to</strong> flip all colored cards and read the<br />

word in the boxes. The robot then thanks the participant and<br />

drives <strong>to</strong> the table on the left-hand side of the participant.<br />

The third part of the experiment is very small. When the robot<br />

arrives at the table it turns around and asks the participant <strong>to</strong><br />

come closer. This part is also needed for the other experiment.<br />

The robot tells the participant there is a questionnaire on the<br />

table next <strong>to</strong> her and asks the participant <strong>to</strong> fill it in. The<br />

experiment has ended.<br />

The figure below shows a picture taken from the experiment<br />

setup. The table with the six squares is in front of a platform<br />

on which the robot drives around. This was necessary because<br />

otherwise the robot was <strong>to</strong>o small and the participant would<br />

not have been able <strong>to</strong> see the robot’s eyes.<br />

The moving assignments are similar <strong>to</strong> the ones used by<br />

Staudte et al. [14]. However this experiment is not based on<br />

utterance or gesture, but instead focuses on the two different<br />

voices.<br />

Figure 2: the experiment setup.<br />

3.4 Questionnaire & video footage<br />

A three-page questionnaire was used <strong>to</strong> determine if a person<br />

was engaged or interested in the robot or task. The<br />

questionnaire is based on measurement instruments for<br />

measuring anthropomorphism, animacy, likeability, perceived<br />

intelligence and perceived safety of robots provided by<br />

Bartneck et al. [2] <strong>to</strong> get understanding of how the robot was<br />

perceived by the participants. The questionnaire also contains<br />

questions on the attention allocation scale provided by Harms<br />

and Biocca [6] <strong>to</strong> find out if the participants were interested in<br />

the robot or the task.<br />

The questionnaire consists of five parts. The first part is about<br />

some general information. The second, third and fifth part<br />

contains 32 questions about the experiment on a 5-point<br />

Likert scale provided by Bartneck et al. [2]. The fourth part<br />

contains 9 questions which are rated on a scale from 1 <strong>to</strong> 7, of<br />

which some are provided by Harms and Biocca [6]. Because<br />

the conducted experiment is shared with Rens Hoegen not all<br />

questions will be used for providing answers in this paper.<br />

3.4.1 Most important questions<br />

The categories ‘rating of the participant’s’ have some<br />

important questions in order <strong>to</strong> answer the research questions:<br />

- artificial / lifelike & mechanical / Organic:<br />

questions in the categories anthropomorphism and<br />

animacy <strong>to</strong> determine how the robot's appearance<br />

was perceived.<br />

- unfriendly / friendly & unpleasant / pleasant:<br />

questions in the category likeability <strong>to</strong> rate the<br />

impression of the robot.<br />

- incompetent / competent & unintelligent /<br />

intelligent: questions in the category perceived<br />

intelligence.<br />

Another category contains questions <strong>to</strong> which the participant<br />

can disagree or agree. The most important are:<br />

- I was interested in the robot.<br />

- I was interested in the task.<br />

- I remained focused on the robot throughout our<br />

interaction.<br />

- I remained focused on the task throughout our<br />

interaction.<br />

The questionnaire ends with a manipulation check <strong>to</strong><br />

determine if the difference in voices was noticed. The<br />

experiment was also recorded by two different cameras (one<br />

full HD camcorder and one fisheye camera) in order <strong>to</strong><br />

support the questionnaire. Some small screenshots taken from<br />

the video footage can be found in section 4.2. The questions<br />

of the questionnaire can be found in Appendix A.<br />

4. RESULTS<br />

Most of the 41 questions from the questionnaire were used <strong>to</strong><br />

analyze the relationship between voices and the level of<br />

interest. Some questions were not relevant for this research<br />

because it was a combined experiment. At first it is important<br />

<strong>to</strong> determine the reliability of the questions, after that the<br />

questions will be used <strong>to</strong> answer the research questions.<br />

Finally some screenshots of the video footage is shown <strong>to</strong><br />

highlight some details.<br />

Figure 3: a schematic representation of the experiment.


4.1 Questionnaire<br />

The categories anthropomorphism, animacy, likeability,<br />

perceived intelligence and perceived safety suggested by<br />

Bartneck et al. [2] were all used in the questionnaire.<br />

4.1.1 Reliability<br />

The Cronbach Alpha value of each of the categories is listed<br />

in the third column of the table below. The second column<br />

shows which questions were used in that particularly<br />

category. The matching questions can be found in Appendix<br />

A.<br />

Table 1: Cronbach Alpha values for each category before<br />

removing questions.<br />

Category Questions (#) Cronbach Alpha<br />

Anthropomorphism 1, 5, 9, 13, 17 0.62<br />

Animacy 2, 6, 10, 13, 14,<br />

18<br />

0.62<br />

Likeability 3, 7, 11, 15, 19 0.81<br />

Perc. intelligence 4, 8, 12, 16, 20 0.61<br />

Perc. safety 24, 25, 26 0.67<br />

Because most Cronbach Alpha values were not very reliable<br />

some questions have been deleted. To achieve a Cronbach<br />

Alpha value of at least 0.70 in each category the following<br />

questions have been deleted:<br />

- moving rigidly / moving elegantly & unconscious /<br />

conscious in the category anthropomorphism.<br />

- apathetic / responsive, stagnant / lively & inert /<br />

interactive in the category animacy.<br />

- foolish / sensible in the category perceived<br />

intelligence.<br />

- quiescent / surprised in the category perceived<br />

safety.<br />

It is important <strong>to</strong> state that some participants had trouble with<br />

the meaning of apathetic and quiescent, all participants were<br />

Dutch.<br />

This results in the following Cronbach Alpha values:<br />

Table 2: Cronbach Alpha values for each category after<br />

removing questions.<br />

Category Questions (#) Cronbach Alpha<br />

Anthropomorphism 1, 5, 13 0.78<br />

Animacy 2, 10, 13 0.72<br />

Likeability 3, 7, 11, 15, 19 0.81<br />

Perc. intelligence 4, 8, 12, 16 0.72<br />

Perc. safety 24, 25 0.85<br />

4.1.2 Analysis<br />

In the tables below the results of the Independent Samples T-<br />

tests is shown. In table 3 the results of the test are shown for<br />

Bartneck’s categories. The results of the questions used in the<br />

manipulation check are shown in table 4.<br />

The other questions, including the questions taken from<br />

Harms and Biocca are evaluated separately in tables 5 and 6.<br />

Table 3: Independent Samples T-tests results for<br />

Bartneck’s categories with both conditions.<br />

Category <strong>Voice</strong> M SD t(19) p<br />

Anthropomorphism<br />

Animacy<br />

Likeability<br />

Perceived<br />

intelligence<br />

Perceived<br />

safety<br />

<strong>Human</strong>-like 2.47 0.48<br />

Machine-like 1.97 0.64<br />

<strong>Human</strong>-like 2.23 0.61<br />

Machine-like 2.03 0.71<br />

<strong>Human</strong>-like 4.00 0.46<br />

Machine-like 3.51 0.47<br />

<strong>Human</strong>-like 3.48 0.38<br />

Machine-like 3.20 0.65<br />

<strong>Human</strong>-like 3.75 0.83<br />

Machine-like 3.86 0.60<br />

2.00 0.060<br />

0.70 0.491<br />

2.42 0.026<br />

1.15 0.266<br />

-0.37 0.719<br />

Analyzing and comparing the two tables above shows that the<br />

categories animacy, perceived intelligence and perceived<br />

safety have no significant difference between the human-like<br />

voice and the machine-like voice. The category likeability on<br />

the contrary does show a significant difference between the<br />

human-like voice and the machine-like voice. The category<br />

anthropomorphism was approaching significance.<br />

Table 4: Independent Samples T-tests results for the<br />

manipulation check with both conditions.<br />

Question <strong>Voice</strong> M SD T(19) p<br />

Machinelike<br />

/ <strong>Human</strong>like<br />

Unpleasant<br />

/ Pleasant<br />

Disengaging<br />

/ <strong>Engaging</strong><br />

Unclear<br />

/ Clear<br />

<strong>Human</strong>-like 2.00 1.05<br />

Machine-like 1.64 0.51<br />

<strong>Human</strong>-like 3.20 0.92<br />

Machine-like 3.00 0.78<br />

<strong>Human</strong>-like 2.90 0.57<br />

Machine-like 3.09 0.70<br />

<strong>Human</strong>-like 3.20 0.92<br />

Machine-like 2.36 0.67<br />

1.02 0.319<br />

0.54 0.595<br />

-0.68 0.504<br />

2.39 0.027<br />

The questions machinelike / humanlike, unpleasant / pleasant<br />

& disengaging / engaging in the manipulation check about the<br />

voice only showed minor differences, mostly in favor of the<br />

human-like model. The most important significant difference<br />

can be found in the question unclear / clear. It showed that the<br />

human-like voice was more clear than the machine-like voice<br />

and that a difference in voices was noticed. This can also be<br />

seen in the video footage, discussed in section 4.2 below.<br />

Some 7-point scale questions are excluded in this research, the<br />

others are listed below in tables 5 and 6. For each of the<br />

questions the Independent Samples T-test is performed and<br />

listed below.<br />

Table 5: Independent Samples T-tests results for the<br />

remaining questions with the human-like voice.<br />

Question<br />

M SD t(19) p<br />

I feel that the robot is interesting <strong>to</strong> look at.<br />

4.80 1.03 1.65 0.115


I was interested in the robot.<br />

5.70 0.48 1.13 0.273<br />

I was interested in the task.<br />

4.70 1.57 0.48 0.636<br />

I was easily distracted from the robot when other things were<br />

going on. (Recoded)<br />

5.40 1.17 1.11 0.281<br />

I remained focused on the robot throughout our interaction.<br />

4.90 1.52 -0.95 0.352<br />

I remained focused on the task throughout our interaction.<br />

5.10 1.20 -1.66 0.112<br />

Understanding the robot was difficult. (Recoded *)<br />

5.50 1.18 1.86 0.078<br />

voices. These pictures will be used in section 6 <strong>to</strong> support the<br />

conclusion.<br />

Figure 3: the participant is fixating on the robot after<br />

completing an assignment.<br />

Table 6: Independent Samples T-tests results for the<br />

remaining questions with the machine-like voice.<br />

Question<br />

M SD t(19) p<br />

I feel that the robot is interesting <strong>to</strong> look at.<br />

3.82 1.601 1.65 0.115<br />

I was interested in the robot.<br />

5.27 1.104 1.13 0.273<br />

I was interested in the task.<br />

4.36 1.629 0.48 0.636<br />

I was easily distracted from the robot when other things were<br />

going on. (Recoded)<br />

4.73 1.555 1.11 0.281<br />

I remained focused on the robot throughout our interaction.<br />

5.45 1.128 -0.95 0.352<br />

I remained focused on the task throughout our interaction.<br />

5.82 0.751 -1.66 0.112<br />

Understanding the robot was difficult. (Recoded *)<br />

4.36 1.567 1.86 0.078<br />

Figure 4: her facial expressions indicate that she has<br />

difficulties with understanding the robot.<br />

The above tables show that the differences between the two<br />

voices in terms of interest is not very significant. Thought it is<br />

interesting <strong>to</strong> know that there is a slight difference in focusing<br />

on either the task or the robot.<br />

* Note that some questions are recoded <strong>to</strong> make sure that all<br />

questions are rated the same way, meaning that 1 = negative,<br />

4 = neutral and 7 = positive. In the question ‘Understanding<br />

the robot was difficult.’ 1 was positive (not difficult) and 7<br />

was negative (difficult), therefore the question is recoded <strong>to</strong><br />

match its rating with the other questions.<br />

4.2 Video footage<br />

In the previous section it became clear that the human-like<br />

voice was more likeable and that understanding the robot was<br />

slightly more difficult with the machine-like voice. The<br />

machine-like voice was rated as more unclear compared <strong>to</strong> the<br />

human-like voice. The video footage showed the exact same<br />

results. The pictures below show some information about the<br />

Figure 5: the participant is staring at the background,<br />

thinking about what the robot just said.<br />

5. DISCUSSION<br />

The difference between the two used voices turned out <strong>to</strong> be<br />

not very significant. The most important differences were<br />

found in Bartneck's category likeability and the question<br />

unclear / clear in the manipulation check. The human-like<br />

voice scored higher on the likeability scale and was found<br />

more clear than the machine-like voice. The questions in the<br />

category anthropomorphism were only approaching<br />

significance and showed that the human-like voice was<br />

perceived as slightly more anthropomorphic. There was also a<br />

minor difference in understanding the robot. The machine-like<br />

voice was harder <strong>to</strong> understand than the human-like voice,<br />

which relates <strong>to</strong> the difference found in the unclear / clear<br />

question.


The relationship between voice and interest is not significant.<br />

In the introduction the question 'If a robot is perceived as<br />

artificial and uses a mechanical voice instead of the usual<br />

human-like voice, does this influence the way humans are<br />

engaged <strong>to</strong> perform a certain task' came <strong>to</strong> mind. The results<br />

show that the influence of the voices is minimal. The robot<br />

with the human-like voice was perceived only a little more<br />

interesting.<br />

The results in table 5 and 6 appear <strong>to</strong> show a contradiction.<br />

The human-model scored slightly better on the questions<br />

about interest. However, the questions about focus show a<br />

minor difference in favor of the machine-like voice.<br />

6. CONCLUSION<br />

As mentioned above, some results appear <strong>to</strong> be contradicting.<br />

Though these results are not as significant as hoped, the<br />

difference can still be explained. Because the machine-like<br />

voice was rated less clear than the human-like voice, some<br />

participants had difficulties with understanding the robot.<br />

Section 4.2 shows three pictures taken from the video footage.<br />

Figure 4 shows a participant with facial expressions indicating<br />

it takes some effort <strong>to</strong> understand the robot. Figure 5 shows a<br />

participant thinking about what the robot said. It seems that<br />

the participants with the machine-like voice focused more on<br />

the robot because of the unclear voice.<br />

With the above results, the research questions mentioned in<br />

section 1.2 can be answered:<br />

This research showed that a robot's voice does not<br />

significantly influence the way participants are engaged <strong>to</strong><br />

perform a certain task. Nor were the participants significantly<br />

more interested in the robot or task.<br />

The human-like voice in general scored overall better than the<br />

machine-like voice. This result is similar <strong>to</strong> the results found<br />

by other researchers mentioned in the introduction.<br />

7. FUTURE WORK<br />

In order <strong>to</strong> completely understand the relationship between<br />

voice and engagement, more research needs <strong>to</strong> be conducted.<br />

To improve this research better utilities are necessary. The<br />

used Magabot did not have a body (or body movement) and<br />

lacked facial expressions. To improve this experiment some<br />

aspects of the robot such as embodiment and perceived gender<br />

(more than just the voice of the robot) should be implemented.<br />

The robot was not quite socially interactive.<br />

The video footage could have been of greater importance <strong>to</strong><br />

this research. Due <strong>to</strong> time issues, it was only possible <strong>to</strong> use<br />

the video footage as support <strong>to</strong> the questionnaire. The analysis<br />

of the video footage might reveal more interesting facts in a<br />

larger study.<br />

Some participants asked how the robot knew they were<br />

finished with a certain task. Parts of the experiment were<br />

conducted using the Wizard of Oz technique, which might<br />

have influenced the way the participant perceived the robot. In<br />

the future it is necessary <strong>to</strong> create a robot which responds<br />

au<strong>to</strong>nomous based on the actions of the participant.<br />

Future work on this research might use a larger group of<br />

participants with equally mixed gender and a less technical<br />

background. The experiment was also conducted in a taskbased<br />

environment, experimenting in a more real-life situation<br />

would be more appropriate.<br />

8. ACKNOWLEDGEMENTS<br />

I would like <strong>to</strong> thank Betsy van Dijk and Manja Lohse for<br />

providing guidance and solving problems during the research.<br />

And I would also like <strong>to</strong> thank the University of Twente for<br />

providing all necessary utilities <strong>to</strong> conduct the experiments.<br />

My thanks also go <strong>to</strong> Gilber<strong>to</strong> Sepúlveda Bradford and Rens<br />

Hoegen for helping with the robot and the experiments.<br />

9. REFERENCES<br />

[1] C. Bartneck and J. Forlizzi. A Design-Centred<br />

Framework for Social <strong>Human</strong>-<strong>Robot</strong> <strong>Interaction</strong>. In<br />

Proceedings of the Ro-Man2004, Kurashiki pp. 591-594.<br />

2004. ACM.<br />

DOI: http://dx.doi.org/10.1109/ROMAN.2004.1374827<br />

[2] C. Bartneck, D. Kulić, E. Croft and S. Zoghbi.<br />

Measurement Instruments for the Anthropomorphism,<br />

Animacy, Likeability, Perceived Intelligence, and<br />

Perceived Safety of <strong>Robot</strong>s. In International Journal of<br />

Social <strong>Robot</strong>ics, Vol 1(1), 71-81. 2009.<br />

DOI=http://dx.doi.org/10.1007/s12369-008-0001-3<br />

[3] C.R. Crowelly, M. Villanoy, M. Scheutz and P.<br />

Schermerhorn. Gendered voice and robot entities:<br />

Perceptions and reactions of male and female subjects. In<br />

Intelligent <strong>Robot</strong>s and Systems, IROS 2009, IEEE/RSJ<br />

International Conference, 3735-3741. 2009.<br />

DOI=http://dx.doi.org/10.1109/IROS.2009.5354204<br />

[4] J. Goetz, S. Kiesler, and A. Powers. Matching robot<br />

appearance and behavior <strong>to</strong> tasks <strong>to</strong> improve humanrobot<br />

cooperation. In Proceedings of ROMAN 2003, the<br />

12th IEEE International Workshop on <strong>Robot</strong> and <strong>Human</strong><br />

Interactive Communication, 55-60, 2003. ACM.<br />

DOI= http://dx.doi.org/10.1109/ROMAN.2003.1251796<br />

[5] L. Gong and C. Nass. When a talking-face computer<br />

agent is half-human and half-humanoid: human identity<br />

and consistency preference. In Journal of <strong>Human</strong><br />

Communication Research, Vol 33(2), 163–193. 2007.<br />

DOI=http://dx.doi.org/10.1111/j.1468-958.2007.00295.x<br />

[6] C. Harms and F. A.Biocca. Internal consistency and<br />

reliability of the networked minds social presence<br />

measure. In M. Alcaniz & B. Rey (Eds.), Seventh Annual<br />

International Workshop. 2004. ACM.<br />

DOI=http://cogprints.org/7026/<br />

[7] K. M. Lee, W. Peng, S. Jin and C. Yan. Can robots<br />

manifest personality An empirical test of personality<br />

recognition, social responses, and social presence in<br />

human-robot interaction. In Journal of Communication,<br />

56, 754-772. 2006.<br />

DOI=http://dx.doi.org/10.1111/j.1460-466.2006.00318.x<br />

[8] J. A. LePine and L. Van Dyne. <strong>Voice</strong> and cooperative<br />

behavior as contrasting forms of contextual performance:<br />

Evidence of differential relationships with Big Five<br />

personality characteristics and cognitive ability. In<br />

Journal of Applied Psychology, Vol 86(2), 326-336.<br />

2001.<br />

DOI=http://psycnet.apa.org/doi/10.1037/0021-<br />

9010.86.2.326<br />

[9] K. MacDorman and H. Ishiguro. The uncanny advantage<br />

of using androids in cognitive and social science<br />

research. <strong>Interaction</strong> Studies, Vol 7(3), 297–337. 2006.<br />

DOI=http://dx.doi.org/10.1075/is.7.3.03mac


[10] B. Mutlu, J. Forlizzi and J. Hodgins. A S<strong>to</strong>rytelling<br />

<strong>Robot</strong>: Modeling and Evaluation of <strong>Human</strong>-like Gaze<br />

Behavior. In Proceedings of HUMANOIDS’06, 2006.<br />

ACM.<br />

DOI= http://dx.doi.org/10.1109/ICHR.2006.321322<br />

[11] Y. Okuno, T. Kanda, M. Imai, H. Ishiguro, and N.<br />

Hagita. Providing route directions: design of robot’s<br />

utterance, gesture, and timing. In Proceedings of the 4th<br />

ACM/IEEE international conference on <strong>Human</strong> robot<br />

interaction, HRI’09, 53-60, 2009. ACM.<br />

DOI=http://doi.acm.org/10.1145/1514095.1514108<br />

[12] C. Peters, G. Castellano and S. de Freitas. An exploration<br />

of user engagement in HCI. In Proceedings of the<br />

affective-aware virtual agents and social robots<br />

(AFFINE) workshop, international conference on<br />

multimodal interfaces, ICMI’09, 2009. ACM.<br />

DOI=http://doi.acm.org/10.1145/1655260.1655269<br />

[13] M. Siegel, C. Breazeal and M. I. Nor<strong>to</strong>n. Persuasive<br />

<strong>Robot</strong>ics: The influence of robot gender on human<br />

behavior. In Intelligent <strong>Robot</strong>s and Systems, IROS<br />

(2009). 2563-2568. 2009. ACM.<br />

DOI=http://dx.doi.org/10.1109/IROS.2009.5354116<br />

[14] M. Staudte and M. Crocker. Visual attention in spoken<br />

human-robot interaction. In Proceedings of the 4th<br />

ACM/IEEE international conference on <strong>Human</strong> robot<br />

interaction, HRI’09, 77–84, 2009. ACM.<br />

DOI=http://doi.acm.org/10.1145/1514095.1514111<br />

[15] A. Steinfeld, T. Fong, D. Kaber, M. Lewis, J. Scholtz, A.<br />

Schultz and M. Goodrich. Common metrics for humanrobot<br />

interaction. In Proceedings of the 1st ACM/IEEE<br />

International Conference on <strong>Human</strong><strong>Robot</strong> <strong>Interaction</strong>s,<br />

HRI'06, 2006. ACM.<br />

DOI= http://dx.doi.org/10.1145/1121241.1121249<br />

[16] M. L. Walters, K. Dautenhahn, R. te Boekhorst, K. L.<br />

Koay, C. Kaouri, S. Woods, C. Nehaniv, D. Lee, and I.<br />

Werry. The influence of subjects’ personality traits on<br />

personal spatial zones in a human-robot interaction<br />

experiment. In Proceedings of COG SCI 2005: Toward<br />

Social Mechanisms of Android Science Workshop, 29-37,<br />

2005. ACM.<br />

DOI= http://dx.doi.org/10.1109/ROMAN.2005.1513803<br />

[17] M. L. Walters, D. S. Syrdal, K. Dautenhahn, R. te<br />

Boekhorst and K. L. Koay. Avoiding the uncanny valley:<br />

robot appearance, personality and consistency of<br />

behavior in an attention-seeking home scenario for a<br />

robot companion. Au<strong>to</strong>nomous <strong>Robot</strong>s, 159-178. 2008.<br />

DOI=http://dx.doi.org/10.1007/s10514-007-9058-3<br />

[18] M. L. Walters, D. S. Syrdal, K. L. Koay, K. Dautenhahn,<br />

R. te Boekhorst. <strong>Human</strong> approach distances <strong>to</strong> a<br />

mechanical-looking robot with different robot voice<br />

styles. In the 17th IEEE International Symposium on<br />

<strong>Robot</strong> and <strong>Human</strong> Interactive Communication, RO-<br />

MAN (2008), 707-712. 2008.<br />

DOI=http://dx.doi.org/10.1109/ROMAN.2008.4600750


APPENDIX<br />

A. RESULTS QUESTIONNAIRE<br />

Table A1: first part of the questionnaire containing questions rated on a scale from 1 <strong>to</strong> 5.<br />

<strong>Human</strong>-like voice<br />

Machine-like voice<br />

# Questions Mean Std. Deviation Mean Std. Deviation<br />

Rate the impression of the robot on a scale from 1 <strong>to</strong> 5.<br />

1 Fake / Natural 3.20 0.632 2.45 0.820<br />

2 Dead / Alive 2.90 1.101 2.55 1.036<br />

3 Dislike / Like 4.00 0.667 3.27 0.647<br />

4 Incompetent / Competent 3.50 0.850 3.55 0.934<br />

5 Machinelike / <strong>Human</strong>like 1.90 0.738 1.64 0.674<br />

6 Stagnant / Lively 3.10 1.370 2.45 0.820<br />

7 Unfriendly / Friendly 4.30 0.675 3.73 0.467<br />

8 Ignorant / Knowledgeable 3.40 0.699 3.09 0.701<br />

9 Unconscious / Conscious 3.10 0.876 2.82 0.874<br />

10 Mechanical / Organic 1.50 0.707 1.73 0.647<br />

11 Unkind / Kind 4.10 0.738 3.55 0.522<br />

12 Irresponsible / Responsible 3.40 0.516 3.00 0.632<br />

13 Artificial / Lifelike 2.30 0.483 1.82 0.751<br />

14 Inert / Interactive 3.60 0.699 3.09 0.831<br />

15 Unpleasant / Pleasant 3.70 0.675 3.45 0.688<br />

16 Unintelligent / Intelligent 3.60 0.516 3.18 0.874<br />

17 Moving rigidly / Moving elegantly 2.80 0.919 2.27 1.009<br />

18 Apathetic / Responsive 3.50 0.850 3.36 0.674<br />

19 Awful / Nice 3.90 0.738 3.55 0.688<br />

20 Foolish / Sensible 3.00 0.471 3.55 0.688<br />

21 Quiet / Loud 2.60 0.699 2.64 0.674<br />

22 Unhelpful / Helpful 3.50 0.527 3.45 0.820<br />

23 Intimidating / Inviting 3.60 0.699 3.55 0.820<br />

Rate your emotional state on a scale from 1 <strong>to</strong> 5.<br />

24 Anxious / Relaxed 3.70 0.949 3.91 0.539<br />

25 Agitated / Calm 3.80 0.789 3.82 0.751<br />

26 Quiescent / Surprised 2.50 0.850 3.00 0.894<br />

27 Unsafe / Safe 4.00 1.054 4.27 0.647<br />

28 Pressured / At ease 3.70 0.823 3.45 0.688


Table A2: second part of the questionnaire containing questions rated on a scale from 1 <strong>to</strong> 7.<br />

<strong>Human</strong>-like voice<br />

Machine-like voice<br />

# Questions Mean Std. Deviation Mean Std. Deviation<br />

Give your opinion on the following statements. (Scale from 1 <strong>to</strong> 7, strongly disagree <strong>to</strong> strongly agree.)<br />

29 I feel that the robot is interesting <strong>to</strong> look at. 4.80 1.033 3.82 1.601<br />

30 I was interested in the robot. 5.70 0.483 5.27 1.104<br />

31 I was interested in the task. 4.70 1.567 4.36 1.629<br />

32 I was easily distracted from the robot when other things<br />

were going on.<br />

33 I remained focused on the robot throughout our<br />

interaction.<br />

34 I remained focused on the task throughout our<br />

interaction.<br />

5.40 1.174 4.73 1.555<br />

4.90 1.524 5.45 1.128<br />

5.10 1.197 5.82 0.751<br />

35 Understanding the robot was difficult. 5.50 1.179 4.36 1.567<br />

36 Throughout our interaction I became more familiar with<br />

the robot.<br />

5.20 0.919 4.45 1.508<br />

37 I felt uncomfortable when I was close <strong>to</strong> the robot. 5.30 0.949 5.55 1.695<br />

Table A3: third part of the questionnaire containing questions rated on a scale from 1 <strong>to</strong> 5, used as the manipulation check.<br />

<strong>Human</strong>-like voice<br />

Machine-like voice<br />

# Questions Mean Std. Deviation Mean Std. Deviation<br />

Rate your impression of the voice of the robot on the following scales.<br />

38 Machinelike / <strong>Human</strong>like 2.00 1.054 1.64 0.505<br />

39 Unpleasant / Pleasant 3.20 0.919 3.00 0.775<br />

40 Disengaging / <strong>Engaging</strong> 2.90 0.568 3.09 0.701<br />

41 Unclear / Clear 3.20 0.919 2.36 0.674

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!