29.12.2012 Views

Prof. Dr. Wolfgang König, Prof. Dr.-Ing. Ralf ... - E-Finance Lab

Prof. Dr. Wolfgang König, Prof. Dr.-Ing. Ralf ... - E-Finance Lab

Prof. Dr. Wolfgang König, Prof. Dr.-Ing. Ralf ... - E-Finance Lab

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Ergebnisband des E-<strong>Finance</strong> <strong>Lab</strong><br />

Top-5-Publikationen je Layer des 1. Halbjahres 2012


Inhaltsverzeichnis<br />

Layer 1: IT-Infrastructure: Service Systems in E-<strong>Finance</strong><br />

(<strong>Prof</strong>. <strong>Dr</strong>. <strong>Wolfgang</strong> <strong>König</strong>, <strong>Prof</strong>. <strong>Dr</strong>.-<strong>Ing</strong>. <strong>Ralf</strong> Steinmetz,<br />

<strong>Prof</strong>. <strong>Dr</strong>. Roman Beck)<br />

� Beck R., Weber S., Gregory R. (2012):<br />

Theory-Generating Design Science Research<br />

In: Information Systems Frontiers - online first, DOI: 10.1007/s10796-012-9342-4.<br />

� Beck R., Schott K. (2012):<br />

The Interplay of Project Control and Interorganizational Learning: Mitigating Effects<br />

on Cultural Differences in Global, Multisource ISD Outsourcing Projects<br />

In: Business & Information Systems Engineering - online first, DOI:10.1007/s12599-<br />

012-0217-5.<br />

� Lampe U., Siebenhaar M., Papageorgiou A., Schuller D., Steinmetz R. (2012):<br />

Maximizing Cloud Provider <strong>Prof</strong>it from Equilibrium Price Auctions<br />

In: Proceedings of the 5 th International Conference on Cloud Computing (CLOUD<br />

2012), Honolulu, Hawaii, USA.<br />

� Schuller D., Lampe U., Schulte S., Eckert J., Steinmetz R. (2012):<br />

Cost-driven Optimization of Complex Service-based Workflows for Stochastic QoS<br />

Parameters<br />

In: Proceedings of the 9 th International Conference on Web Services (ICWS 2012),<br />

Honolulu, Hawaii, USA.<br />

� Wolf M., Beck R., Pahlke I. (2012):<br />

Mindfully Resisting the Bandwagon – Reconceptualising IT Innovation Assimilation in<br />

Highly Turbulent Environments<br />

In: Journal of Information Technology (JIT) - Online first, DOI:10.1057/jit.2012.13.<br />

Layer 2: E-Financial Markets & Market Infrastructures<br />

(<strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber)<br />

� Gomber P., Pujol G., Wranik A. (2012):<br />

Best Execution Implementation and Broker Policies in Fragmented European Equity<br />

Markets<br />

In: International Review of Business Research Papers, Vol. 8, Issue 2, 144-162.<br />

� Haferkorn M., Lutat M., Zimmermann K. (2012):<br />

The Effect of Single-Stock Circuit Breakers on the Quality of Fragmented Markets<br />

In: <strong>Finance</strong>Com 2012, Barcelona, Spain.<br />

� Lattemann C., Loos P., Johannes G., Burghof H., Breuer A., Gomber P.,<br />

Krogmann M., Nagel J., Riess R., Riordan R., Zajonz R. (2012):<br />

High Frequency Trading - Costs and Benefits in Securities Trading and its Necessity<br />

of Regulations<br />

In: Business & Information Systems Engineering, Vol. 4, Issue 2, 93-108.


� Siering, M. (2012):<br />

Investigating the Market Impact of Media Sentiment and Investor Attention<br />

In: <strong>Finance</strong>Com 2012; Barcelona, Spain.<br />

� Weber M.C., Wondrak C. (2012):<br />

Measuring the Influence of Project Characteristics on Optimal Software Project<br />

Granularity<br />

In: Proceedings of the 20 th European Conference on Information Systems (ECIS),<br />

Barcelona, Spain.<br />

Layer 3: Customer Management in E-<strong>Finance</strong><br />

(<strong>Prof</strong>. <strong>Dr</strong>. Andreas Hackethal, <strong>Prof</strong>. <strong>Dr</strong>. Bernd Skiera, <strong>Prof</strong>.<br />

<strong>Dr</strong>. Oliver Hinz)<br />

� Bhattacharya U., Andreas H., Simon K., Benjamin L., Steffen M. (2012):<br />

Is unbiased financial advice to retail investors sufficient? Answers from a large field<br />

study<br />

In: Review of Financial Studies, Vol. 25, 975-1032.<br />

� Hackethal A., Michael H., Tullio J. (2012):<br />

Financial advisors: A case of babysitters?<br />

In: Journal of Banking & <strong>Finance</strong>, Vol. 36, 509-524.<br />

� Schmitt P., Skiera B., Van den Bulte C. (2011):<br />

Referral Programs and Customer Value<br />

In: Journal of Marketing, Vol. 75, Jan., 46-59.<br />

� Hinz O., Skiera B., Barrot C., Becker J. (2011):<br />

An Empirical Comparison of Seeding Strategies for Viral Marketing<br />

In: Journal of Marketing, Vol. 75, Nov., 55-71.<br />

� Skiera B., Bermes M., Horn L. (2011):<br />

Customer Equity Sustainability Ratio: A New Metric for Assessing a Firm’s Future<br />

Orientation<br />

In: Journal of Marketing, Vol. 75, May, 118-131.<br />

Die Autoren bedanken sich herzlich für die Unterstützung aller ihrer Arbeiten durch<br />

das E-<strong>Finance</strong> <strong>Lab</strong>!


Layer 1:<br />

IT-Infrastructure: Service Systems in E-<strong>Finance</strong><br />

(<strong>Prof</strong>. <strong>Dr</strong>. <strong>Wolfgang</strong> <strong>König</strong>, <strong>Prof</strong>. <strong>Dr</strong>.-<strong>Ing</strong>. <strong>Ralf</strong><br />

Steinmetz, <strong>Prof</strong>. <strong>Dr</strong>. Roman Beck)<br />

� Beck R., Weber S., Gregory R. (2012):<br />

Theory-Generating Design Science Research<br />

In: Information Systems Frontiers - online first.<br />

� Beck R., Schott K. (2012):<br />

The Interplay of Project Control and Interorganizational Learning: Mitigating<br />

Effects on Cultural Differences in Global, Multisource ISD Outsourcing Projects<br />

In: Business & Information Systems Engineering - online first.<br />

� Lampe U., Siebenhaar M., Papageorgiou A., Schuller D., Steinmetz R.<br />

(2012):<br />

Maximizing Cloud Provider <strong>Prof</strong>it from Equilibrium Price Auctions<br />

In: Proceedings of the 5 th International Conference on Cloud Computing<br />

(CLOUD 2012), Honolulu, Hawaii, USA.<br />

� Schuller D., Lampe U., Schulte S., Eckert J., Steinmetz R. (2012):<br />

Cost-driven Optimization of Complex Service-based Workflows for Stochastic<br />

QoS Parameters<br />

In: Proceedings of the 9 th International Conference on Web Services (ICWS<br />

2012), Honolulu, Hawaii, USA.<br />

� Wolf M., Beck R., Pahlke I. (2012):<br />

Mindfully Resisting the Bandwagon – Reconceptualising IT Innovation<br />

Assimilation in Highly Turbulent Environments<br />

In: Journal of Information Technology (JIT) - Online first.


Inf Syst Front<br />

DOI 10.1007/s10796-012-9342-4<br />

Theory-generating design science research<br />

Roman Beck & Sven Weber & Robert Wayne Gregory<br />

# Springer Science+Business Media, LLC 2012<br />

Abstract A frequently mentioned challenge in design<br />

science research (DSR) is the generation of novel theory<br />

above and beyond information technology artefacts.<br />

This article analyzes the DSR process and extends<br />

established frameworks for theory generation to exemplify<br />

improvements to theory generation through methods<br />

of grounded theory development. On a conceptual<br />

base, we developed a theory-generating DSR approach<br />

which integrates methods of grounded theory development<br />

with established DSR methodology. This combination<br />

enables a design theorist to generate theoretical<br />

knowledge that extends the applicable knowledge base.<br />

We do not elaborate this combination on a meta-level,<br />

but rather provide a process model for researchers in<br />

form of an extension of a well-known DSR model to<br />

combine both methods in a pluralistic research design.<br />

With this suggested research approach, scholars can<br />

draw theoretical insights from analytical abstractions<br />

and can improve the development of IT artefacts in a<br />

structured way to avoid failure or repair loops.<br />

R. Beck : S. Weber (*)<br />

Institute of Information Systems, Goethe University Frankfurt,<br />

Grüneburgplatz 1,<br />

60323 Frankfurt am Main, Germany<br />

e-mail: svweber@wiwi.uni-frankfurt.de<br />

R. Beck<br />

e-mail: rbeck@wiwi.uni-frankfurt.de<br />

R. W. Gregory<br />

University of Göttingen,<br />

Platz der Göttinger Sieben 5,<br />

37073 Göttingen, Germany<br />

R. W. Gregory<br />

e-mail: gregory@wiwi.uni-goettingen.de<br />

Keywords Design science research . Behavioural science<br />

research . Grounded theory method . Pluralistic research<br />

method<br />

1 Introduction<br />

In recent years, design science research (DSR) has attracted<br />

more information systems (IS) research attention, because it<br />

deals with the generation of information technology (IT)<br />

artefacts and their evaluation, which is just as important to<br />

IS research as is research into the impacts of IS (Benbasat<br />

and Zmund 2003; Hevneretal.2004). In general, DSR<br />

provides models and guidelines to researchers that enable<br />

them to create, improve, and evaluate IT artefacts (Hevner et<br />

al. 2004; Holmström et al. 2009; March and Smith 1995;<br />

Weber et al. 2012). Yet many existing DSR research<br />

attempts pay less attention in generating an original theoretical<br />

contribution that goes beyond problem-solving IT artefacts<br />

(Carlsson 2006; Gregory and Muntermann 2011;Hevner<br />

et al. 2004; Winter 2008).<br />

We therefore investigate potential avenues for combining<br />

both aims, using a conceptual approach based on an extensive<br />

analysis of DSR and design theory literature. As a result<br />

of our analysis, we propose a theory-generating DSR approach<br />

that is informed by behavioural science elements<br />

used for theorizing, such as techniques from the grounded<br />

theory method (GTM) (e.g., Glaser 1978; Glaser1998;<br />

Glaser and Strauss 1967). Therefore, our proposed approach<br />

closes some methodological gaps in DSR to achieve insights<br />

at a higher level of analytical abstraction (Yadav 2010).<br />

Certain techniques from GTM are remarkably appropriate,<br />

in that they generate additional theoretical insights in DSR.<br />

The proposed theory-generating DSR approach combines


elements from DSR and GTM for simultaneous problem<br />

solving and theory generation.<br />

Calls for more pluralistic research in IS (e.g., Mingers<br />

2001) have prompted some prior attempts to combine different<br />

methods. For example, Goldkuhl (2004) uses behavioural<br />

science techniques in a DSR approach and suggests<br />

that three types of grounding (internal, empirical, and theoretical)<br />

enhance DSR as a means to generate practical<br />

knowledge. Holmström et al. (2009) accumulates DSR with<br />

a second research cycle that features the development of<br />

substantive and formal theory (the focus of GTM) to contribute<br />

to the knowledge base, but they also use behavioural<br />

science elements to develop an explanatory, theory-testing<br />

approach instead of a means to develop new theoretical<br />

insights. Kuechler and Vaishnavi (2008a, b) address theory<br />

development and theorizing in the context of DSR but<br />

without combining any specific behavioural science research<br />

method. A more advanced approach is presented by<br />

Sein et al. (2011), who integrate DSR with action research in<br />

one model. Their model argues that design science needs to<br />

go beyond the traditional focus on the IT artefact and recognize<br />

its organizational embeddedness. By integrating design<br />

science with action research lenses, the authors are able<br />

to emphasize the organizational interventions through IT<br />

artefact deployment as well as the learning from this interventions<br />

in terms of contributions to the knowledge base. In<br />

this critical reflection and learning step, which is somewhat<br />

disconnected from the traditional build and evaluate DSR<br />

cycle, the researchers abstract the key theoretical findings<br />

that go beyond solving an individual problem. Finally, some<br />

other studies combine DSR with action research (e.g., Allen<br />

et al. 2000), ethnography (e.g., Baskerville and Stage 2001),<br />

or soft design science (Baskerville et al. 2009) withthe<br />

goals of improving theoretical abstraction and knowledge<br />

generation within a DSR project.<br />

In summary, existing research frameworks guided towards<br />

the simultaneous use of DSR and behavioural science but fall<br />

short to provide specific processes for how to conduct a multi<br />

method DSR project (Carlsson 2006). Yet both Goldkuhl<br />

(2004) and Holmström et al. (2009) concludethatDSR<br />

and behavioural science may coexist within a single<br />

project but rather as an additional and independent step.<br />

In this paper, we extend this view and develop a process<br />

model for theory-generating DSR that relies on the simultaneous<br />

and not separated use of both DSR and GTM elements.<br />

The structure of the remainder of this article follows the<br />

building blocks of conceptual research described by Yadav<br />

(2010, p. 3). In the next two sections, we provide an overview<br />

of the two research approaches (DSR and GTM)<br />

before we outline how design theory components map onto<br />

the DSR approach. In the subsequent section, we present a<br />

detailed description of the theory-generating DSR approach,<br />

followed by an overview on theory-generating DSR. We<br />

also detail how the IS design theory components relate to<br />

our proposed approach and provide a first illustration how<br />

theory-generating DSR can be applied in DSR projects. We<br />

conclude with a discussion section and an outlook for further<br />

research.<br />

2 Design science research<br />

Inf Syst Front<br />

The primary focus of IS research is the IT artefact (Benbasat<br />

and Zmund 2003), whereas DSR centres on the design and<br />

creation of the artificial (Simon 1969), especially IT artefacts<br />

(in the form of a construct, model, method, instantiation, or<br />

combination thereof; (March and Smith 1995)). Therefore,<br />

DSR encompasses the creation of an innovative construct that<br />

has not existed before and can be used to serve human purposes<br />

(March and Smith 1995). Its historical origins stem<br />

from engineering (Au 2001; March and Smith 1995), but<br />

DSR also relates to several other academic disciplines, such<br />

as architectural science and computer science. A common<br />

ground across these disciplines is a dual focus on both the<br />

practical application of the IT artefact and the scientific abstraction<br />

and learning (Baskerville 2008). In this context,<br />

theorizing in DSR is worthy of discussion and at least as<br />

important as problem-solving itself (Lee et al. 2011).<br />

Most DSR researchers also recognize that the developed<br />

IT artefact can provide theoretical contributions, provided<br />

that key DSR guidelines are followed (Gregor 2002, 2006)<br />

and so-called kernel theories, drawn from the natural and<br />

social sciences through a creative translation process, are<br />

applied (Hevner et al. 2004; Markus et al. 2002; Walls et al.<br />

1992). Derived principles and concepts of such kernel theories<br />

may provide a basis for advancing DSR by specifying<br />

requirements and generating an IT artefact (Walls et al.<br />

1992). However, Gregor and Jones (2007) explorethat<br />

theorization is a key goal in theory-generating DSR which<br />

mayeventuallyleadtoanISdesigntheory.Therefore,<br />

problem solutions and knowledge contributions can be derived<br />

from either existing kernel theories and IT artefacts or<br />

the development of a new IT artefact and subsequent theorization<br />

(Holmström et al. 2009).<br />

In the DSR process, the initial step must be the search for a<br />

problem that has practical relevance (Hevner et al. 2004). In<br />

other words, “a DSR approach seeks a solution to a real-world<br />

problem of interest to practice” (Kuechler and Vaishnavi<br />

2008a, p. 492), which requires a differentiation between products<br />

(IT artefact) and processes (activities that lead to the IT<br />

artefact) in the DSR cycle (March and Smith 1995; Walls et al.<br />

1992). The product outcome is inevitably embedded in<br />

some place, time, and community and must undergo<br />

theorization to meet innovative and progressive demands<br />

(Orlikowski and Iacono 2001).


Inf Syst Front<br />

The DSR process in turn consists of two basic processes:<br />

building and then evaluating the IT artefact (Baskerville et<br />

al. 2009; Hevner and March 2003; March and Smith 1995).<br />

In the building process, a sequence of activities aims to<br />

produce “something new,” then in the evaluation process,<br />

the created IT artefact undergoes evaluation to produce<br />

feedback and generate new knowledge about the problem.<br />

The newly generated insights serve to improve the quality of<br />

the IT artefact and the design process itself (Hevner et al.<br />

2004). These processes take place partly in parallel and<br />

involve multiple iterations, which enables the IT artefact to<br />

be generated such that it fully satisfies the researchers and<br />

practitioners who later make use of it (Markus et al. 2002).<br />

In total then, DSR offers a rigorous and meaningful contribution<br />

to practice and theory, in the form of an IT artefact<br />

and its evaluation (Gregor 2006).<br />

3 Grounded theory method<br />

The supporting component for theory-generation in our<br />

model is GTM. Since it was introduced to sociology by<br />

Glaser and Strauss (1967), the method has been widely<br />

developed and applied in various disciplines (Bryant and<br />

Charmaz 2007a; Urquhart2007). In a grounded theory<br />

study, the focus lies on the discovery or generation of theory<br />

grounded on empirical evidence (Glaser and Strauss 1967).<br />

A grounded theory study considers the process of discovering<br />

concepts and categories and the relationships between<br />

them (Bryant and Charmaz 2007b), with the end result of a<br />

substantive or grounded theory (Glaser 2007).<br />

Over time, GTM has evolved in different directions. For<br />

example, in the mixed methods approach, GTM combines<br />

with other techniques and methods, such as case study<br />

research (Eisenhardt 1989; Eisenhardt and Graebner 2007).<br />

The most predominant approaches are those propagated by<br />

the originators, Glaser and Strauss, who offer “Glaserian”<br />

and “Strausserian” forms of grounded theory (Ketokivi and<br />

Mantere 2010). This distinction also has been called the<br />

emerging versus forcing debate. The proponents of Glaserian<br />

grounded theory place more emphasis on the principle of<br />

emergence; grounded theory should emerge from the data<br />

and existing theories or concepts, and coding schemes<br />

should not be forced onto the data (Udo 2005).<br />

The term “grounded theory” refers to both the end product<br />

and the process. Conducting grounded theory research<br />

involves various techniques prescribed by GTM (Glaser<br />

1998; Strauss and Corbin 1990), including—among others<br />

—theoretical sampling and the constant comparative method<br />

(Suddaby 2006). Theoretical sampling means that<br />

insights from the initial data collection and analysis guide<br />

subsequent data collection and analysis, so the grounded<br />

theory can emerge over time through iterative cycles of<br />

deeply intertwined data collection and analysis. Urquhart<br />

et al. (2010) provide an in-depth discussion of this key<br />

grounded theory concept and refer to this as deciding on<br />

analytic grounds where to sample from next. Over time,<br />

researchers achieve theoretical saturation because additional<br />

data collection and analysis efforts do not yield new findings<br />

(Eisenhardt 1989). In the constant comparative method, the<br />

researcher constantly compares instances of data labelled as<br />

a particular category with other instances of data in the same<br />

category as a means to substantiate these categories and<br />

build theory. This method is discussed at length by Urquhart<br />

et al. (2010) and the authors refer to this as an important tool<br />

for exposing generated analytical insights to rigorous scrutiny<br />

and building theory. Thereby, all kinds of “slices” of data<br />

(e.g., primary data such as qualitative interviews, secondary<br />

data such as documentation and extant literature) are used to<br />

reach higher levels of abstraction and advance conceptualization.<br />

The relations identified among the categories and the<br />

theoretical integration lead to the grounded theory.<br />

Grounded theory method can in principle be used within a<br />

wide range of epistemological stances and research<br />

approaches (Bryant and Charmaz 2007a; Madill et al. 2000).<br />

It provides the ability to generate an in-depth understanding of<br />

a phenomena by analyzing qualitatively the collected data<br />

from different sources, sorting it into consistent categories,<br />

and emerging a grounded insights of it (Urquhart et al. 2010).<br />

For an overview of the process of grounded theory development,<br />

we recommend, for example, Fernandez (2004). In the<br />

past, Weedman (2008) used this multi-epistemological nature<br />

of GTM to analyze an IT artefact design project (called<br />

Sequoia) with elements of GTM. While GTM is never mentioned,<br />

her analyzing techniques were very similar to theoretical<br />

sampling and the constant comparison method. In<br />

addition, Baskerville and Pries-Heje (1999) combinedaction<br />

research with grounded theory to add rigor and reliability to<br />

the theory formulation process of action research. As an<br />

contrary example, Arazy et al. (2010) used qualitative methods<br />

to evaluate and test a social recommender systems that<br />

utilizes data regarding users’ social relationships by filtering<br />

relevant information to users. However, this approach is more<br />

on theory testing and not building. Hence, GTM, as a behavioural<br />

method provides suitable instruments for an in-depth<br />

understanding of the problem space and its environmental<br />

factors to theorize and not only test in DSR.<br />

4 IS design theory and design science research<br />

On the one hand, DSR provides scholarly contributions based<br />

on the design, evaluation, generalization, and/or theorization<br />

of the IT artefact (Gregor 2006; Gregor and Jones 2007;<br />

Orlikowski and Iacono 2001). On the other hand, it provides<br />

a practical contribution to practitioners in the form of the useful


IT artefact (Hevner et al. 2004). Hence, both the IS design<br />

theories as well as extant DSR models have to be taken into<br />

account to develop a new theory-generating DSR approach.<br />

4.1 IS design theory<br />

As a theoretical lens, our literature review focused on fundamental<br />

components of IS design theory. Our initial DSR<br />

understanding pushed us toward Walls et al. (1992) who laid<br />

the basic foundations for developing a precise understanding<br />

about the nature and anatomy of a design theory. They developed<br />

seven components of an IS design theory guided by prior<br />

literature (e.g., Dubin 1978; Nagel1961) and separated them<br />

into product and process components.<br />

The product components encompass meta-requirements,<br />

meta-design, existing kernel theories governing the requirements,<br />

and a test of whether the meta-design satisfies the<br />

meta-requirements. Most existing DSR publications spend<br />

little time considering the requirements or assume they are<br />

already specified (e.g., Aalst and Kumar 2003; Abbasi and<br />

Table 1 Eight components of IS design theory in different DSR approaches<br />

IS Design Theory<br />

Component<br />

Description Comparison with Existing<br />

DSR Frameworks<br />

1. Purpose and scope “What the system is for,” that is, the set<br />

of meta-requirements that specifies the<br />

type of artefact to which the theory<br />

applies and also defines the scope or<br />

boundaries of theory<br />

2. Constructs Representations of entities of interest in<br />

the theory (design sub-parts)<br />

3. Principle of form<br />

and function<br />

The abstract blueprint or architecture<br />

that describes an IT artefact, whether<br />

a product or method/intervention<br />

4. Artefact mutability Changes in the state of the artefact<br />

anticipated by theory; that is, what<br />

degree of artefact change is<br />

encompassed by theory?<br />

5. Testable<br />

propositions<br />

6. Justification<br />

knowledge<br />

7. Principles of<br />

implementation<br />

8. Expository<br />

instantiation<br />

Chen 2008; Umapathy et al. 2008; for further information<br />

please see the Appendix). Because requirement specification<br />

is a central part of developing an IT artefact, as well as<br />

of subsequent satisfaction with its functionality, we regard<br />

this requirements specification phase critical for DSR. In<br />

addition, a lot of information becomes available from a<br />

system when we consider its purpose, the problem it<br />

aims to solve, and the overall intentions behind building<br />

it. Therefore, we searched for an appropriate theoretical<br />

lens to integrate requirements into our theory-generating<br />

DSR approach.<br />

The process components of Walls et al. (1992) represent<br />

the design method for the IT artefact and integrate existing<br />

kernel theories that govern the design process. They also<br />

describe the underlying knowledge for the design process<br />

and guide research projects.<br />

In turn, using the components of design theory identified<br />

by Walls et al. (1992), Gregor and Jones (2007) suggest<br />

eight components that an IS design theory should include.<br />

We build on these eight components to analyze existing<br />

Meta-requirements<br />

(Walls et al. 1992)<br />

Partly covered by the design cycle:<br />

build and evaluate (March and<br />

Smith 1995)<br />

Different frameworks provide<br />

guidelines and recommendations<br />

for the design process (e.g.,<br />

Hevner et al. 2004)<br />

Partly covered by theorizing about<br />

the IT artefact (Orlikowski and<br />

Iacono 2001)<br />

Truth statements about the design theory Evaluation of whether the design<br />

satisfies requirements (Markus<br />

et al. 2002)<br />

The underlying knowledge or theory<br />

from natural or social sciences that<br />

gives a basis and explanation for the<br />

design (kernel theories)<br />

Description of processes for<br />

implementing the theory (either<br />

product or method) in specific<br />

contexts<br />

A physical implementation of the IT<br />

artefact that can help represent the<br />

theory as an expository device and<br />

for the purposes of testing<br />

Underlying kernel theories<br />

(Walls et al. 1992)<br />

Partly covered by the people<br />

involved in a given or changing<br />

environment (Hevner et al. 2004)<br />

A viable IT artefact in the form of<br />

a construct, model, method, or<br />

instantiation (Hevner et al. 2004)<br />

Inf Syst Front<br />

Improvements Required<br />

for Theory-Generating DSR<br />

A more detailed requirements<br />

analysis increases our<br />

fundamental understanding of<br />

the IT artefact and its sub-parts<br />

Detailed theoretical analysis<br />

of the IT artefact and its design<br />

process to adapt to changing<br />

environments and allow for<br />

evolution over time<br />

Evaluation and theory generation<br />

through theory building from<br />

behavioural science (e.g.,<br />

Holmström et al. 2009)<br />

Extend underlying kernel theories<br />

with behavioural science aspects<br />

(Gregor and Jones 2007)<br />

A detailed analysis of influential<br />

factors (e.g., interaction between<br />

involved people)


Inf Syst Front<br />

DSR frameworks and determine if they include all the<br />

components needed to provide theoretical insights (see<br />

Table 1). In the first and second columns of Table 1, we<br />

outline each required design theory component and then<br />

provide information about existing DSR frameworks. Finally,<br />

we illustrate ways to overcome the challenges we have<br />

identified to generate additional theoretical insights from<br />

DSR. Although many IS design theory components have<br />

been addressed, others might be improved or strengthened<br />

by behavioural science elements, particularly to formulate<br />

and evaluate the theoretical insights derived from IT artefact<br />

development and use.<br />

4.2 Extant DSR models<br />

To develop our theory-generating DSR approach, we<br />

searched for a established DSR model that could serve as<br />

a basis for our approach. Several scholarly contributions<br />

focus on the epistemological positioning of DSR. For example,<br />

Winter (2008) differentiates design science from<br />

design research: The former encompasses reflection on and<br />

guidance for the IT artefact construction and evaluation<br />

process while the latter encompasses the creation and evaluation<br />

of a specific IT artefact. Hevner et al. (2004) focus on<br />

the process of IT artefact development. Thus, IS research<br />

appears influenced by both the environment (people,<br />

organizations, existing technology) and the knowledge<br />

base (Hevner et al. 2004). This classification is closely<br />

related to Winter’s (2008) point of view where an IT<br />

artefact results from the constant iteration of refinement<br />

and assessment. In this regard, Hevner et al. (2004), suggest<br />

seven guidelines for conducting DSR, but many other<br />

examples of DSR frameworks and models exist, including<br />

Peffers et al. (2008), Pries-Heje and Baskerville (2008),<br />

March and Smith (1995), and Nunamaker et al. (1991). These<br />

DSR frameworks build the foundation for many researchers<br />

conducting DSR but do not provide explicit prescriptions how<br />

to achieve theoretical abstraction from an IT artefact in a<br />

structured way.<br />

A somewhat different approach is provided by Vaishnavi<br />

and Kuechler (2008) as well as Kuechler and Vaishnavi<br />

(2008a, b) who propose a design cycle for DSR that address<br />

theory development and theorizing. Moreover, they provide<br />

a general model describing each process step in DSR as<br />

illustrated in Fig. 1.<br />

According to the model of Vaishnavi and Kuechler<br />

(2008) as well as Kuechler and Vaishnavi (2008a, b), problem<br />

awareness is the starting point of DSR which is<br />

reflected by an initial proposal depicting the problem that<br />

has to be solved. This step is similar to Hevner et al. (2004)<br />

and their demand for an initial problem that has to be solved.<br />

The next phase is the suggestion phase where it is tested if<br />

the formulated proposal can be transferred into a tentative<br />

Knowledge<br />

Flows<br />

Process<br />

Steps<br />

Awareness of<br />

Problem<br />

Suggestion<br />

Development<br />

Evaluation<br />

Conclusion<br />

Outputs<br />

Proposal<br />

Tentative<br />

Design<br />

Artefact<br />

Performance<br />

Measures<br />

Results<br />

Fig. 1 General design cycle of DSR (Kuechler and Vaishnavi 2008a,<br />

b; Vaishnavi and Kuechler 2008)<br />

design or not. Thereafter, the IT artefact is developed and<br />

evaluated and conclusions are drawn from the IT artefact as<br />

a result of the problem solving process. However, the whole<br />

process is highly repetitive in the sense that every process<br />

step can lead to a repetition and improvement of prior steps<br />

through the knowledge flows depicted in Fig. 1.<br />

Departing from the described DSR model, we used its<br />

basic elements to structure and develop our theorygenerating<br />

DSR approach which incorporates the findings<br />

from Table 1, as illustrated in the following.<br />

5 Theory-generating design science research<br />

On the basis of prior findings and identified components in<br />

DSR design theory development, as well as the ways they<br />

might be improved with behavioural science elements, we<br />

developed our theory-generating DSR approach, which<br />

combines DSR and GTM techniques, as illustrated in<br />

Fig. 2. The proposed approach extends the general design<br />

cycle of DSR from Kuechler and Vaishnavi (2008a, b).<br />

Therefore, the first steps encompass typical methods used<br />

already in DSR: the awareness of the problem as well as to<br />

develop and evaluate the IT artefact. These steps are complemented<br />

and specified by supportive GTM techniques,<br />

such as detailed analyses of the tentative design and requirements,<br />

theoretical sampling, and the constant comparative<br />

technique. A subsequent theory-generation step which is an<br />

extension of Kuechler and Vaishnavi’s (2008a, b) model<br />

encompass the simultaneous use of DSR and GTM to<br />

generate additional theoretical insights, such as by collecting<br />

additional data from the application of the IT artefact in<br />

an appropriate environment. A complementary research<br />

approach can create new insights and results in the conclusion<br />

step to be integrated into existing knowledge, in<br />

addition to the developed IT artefact. The loop back to<br />

the awareness of the problem to refine the tentative design


Fig. 2 Process of generating<br />

additional scholarly<br />

contributions by theorizing in<br />

DSR<br />

and the requirements therefore closes. Moreover, these new<br />

insights and knowledge provide a theoretical basis for<br />

upcoming DSR projects.<br />

In the following sections, we discuss in detail the different<br />

steps of our proposed theory-generating DSR approach<br />

(Fig. 2).<br />

5.1 Awareness of problem and suggestion<br />

The awareness of the problem and the initial definition of<br />

the research lens is critical, in the sense that the underlying<br />

theories of different research approaches (e.g., design or<br />

behavioural science) justify and explain the researcher’s<br />

design process (Gregor and Jones 2007; Kuechler and<br />

Vaishnavi 2008a, b). Unfortunately, serious methodological<br />

problems can occur in this step, for example, a researcher<br />

with a strict GTM lens could primarily focus on exploring<br />

the theoretical implications behind the requirements for the<br />

IT artefact and pay less attention to problem-solving. In<br />

contrast, a DSR researcher might deemphasize important<br />

behavioural issues by focusing only on the actual problem.<br />

Theory-generating DSR therefore tries to address both practical<br />

relevance and theoretical contributions. In contrast to<br />

Kuechler and Vaishnavi's (2008a, b) model, we merged the<br />

awareness of the problem and the suggestion in one step. As<br />

they mention themselves, both complement each other and<br />

are highly intertwined (Kuechler and Vaishnavi 2008a, b).<br />

Hence, both can be combined with our theory-generating<br />

approach in one step to explore the requirements of the IT<br />

artefact.<br />

5.1.1 Theoretical sampling<br />

Inf Syst Front<br />

A first tentative design of the IT artefact and well-defined<br />

requirements are key for problem solving, so this step is<br />

heavily influenced by the real-world problem as it tries to<br />

determine the proper requirements and entities of interest<br />

with regard to the IT artefact (Gregor and Jones 2007).<br />

Researchers therefore need to focus on the actual problem,<br />

but they cannot lose perspective on potentially emergent<br />

problems and must actively integrate solutions into the<br />

development of the IT artefact. By focusing on both goals<br />

simultaneously, researchers can establish a basis for a satisfying<br />

IT artefact. Theoretical sampling in GTM means that<br />

insights from initial data collection and analysis efforts<br />

guide subsequent ones. In other words, understanding of<br />

the phenomena (or business requirements) emerges over<br />

time through iterative cycles of data collection and analysis


Inf Syst Front<br />

that are deeply intertwined (Glaser 1978). Urquhart et al.<br />

(2010) define theoretical sampling as “deciding on analytic<br />

grounds where to sample from next” (p. 371). Thereby,<br />

theoretical sampling assists the classification of data, the characterization<br />

of relationships between data, and to clarify these<br />

relationships. Without this method, it is nearly impossible for<br />

the researcher to generate theoretical insights about the phenomenon<br />

(Urquhart et al. 2010). Accordingly, subsequent data<br />

collection and analysis efforts over time must build upon one<br />

another for cumulative generation of theoretical insights. Over<br />

time, researchers reach a kind of saturation, such that additional<br />

data collection and analysis efforts do not yield any new<br />

findings (Eisenhardt 1989). Moreover, theoretical sampling<br />

increases the researchers’ flexible ability to look for new<br />

aspects that emerge during the process. It also provides a<br />

means to consider potential problems with the IT artefact in<br />

advance, thus enhancing the likelihood of creating innovative<br />

and progressive IT artefacts for testing.<br />

5.1.2 Slices of data<br />

The slices of data to be collected and analyzed relate to the<br />

environment, including people, the organization, and technology,<br />

as well as the knowledge base or extant literature<br />

(Hevner et al. 2004). However, in DSR, slices of data<br />

mainly refer to extant literature or models (e.g., Aalst and<br />

Kumar 2003), collected data (e.g., Albert et al. 2004), or<br />

kernel theories (e.g., Markus et al. 2002) (for a detailed<br />

discussion, see the Appendix). The inclusion of extant literature<br />

alongside the empirical data is a technique widely<br />

adopted by IS grounded theorists (e.g., Fernandez 2004;<br />

Levina and Vaast 2005) as a means to raise the overall<br />

analysis to a higher conceptual level. In addition, the process<br />

of systematic data collection and analysis helps to<br />

identify and clarify the requirements and entities of interest<br />

(Gregor and Jones 2007), which provide the basis for the<br />

subsequent development or adaptation of an IT artefact. This<br />

process is inherently iterative, in that data collection and<br />

analysis on the one hand and IT artefact creation and evaluation<br />

on the other hand are deeply intertwined.<br />

5.1.3 Tentative design and requirements<br />

A first tentative design and well-defined requirements are<br />

extremely important to DSR, because they ensure the IT<br />

artefact’s relevance to a real-world problem (Gregor and<br />

Jones 2007; Hevner et al. 2004; Pries-Heje and Baskerville<br />

2008; Walls et al. 1992). In our theory-generating DSR<br />

approach, these requirements and entities emerge from the<br />

intertwined process of theoretical sampling and the collection<br />

of slices of data. Most DSR projects integrate the IT<br />

artefact’s requirements as an essential step (e.g., Aalst and<br />

Kumar 2003; Albert et al. 2004; Markus et al. 2002; for a<br />

detailed discussion please see the Appendix), because without<br />

clearly defined requirements, the IT artefact will not be<br />

useful and cannot offer a satisfying solution.<br />

5.2 Development<br />

The tentative design and the explored requirements lead to<br />

the development and creation of the IT artefact which is the<br />

outcome of this step (Kuechler and Vaishnavi 2008a, b).<br />

Because DSR has its roots in engineering and the science of<br />

the artificial (Au 2001; Baskerville 2008; Hevneretal.<br />

2004; McKay and Marshall 2005), the creation and evaluation<br />

of the IT artefact are frequently at the focus of researcher’s<br />

attention. They must exist in any theory-generating<br />

DSR approach. In addition, DSR provides additional theoretical<br />

contributions by representing both an expository<br />

device and the purpose of the testing (Gregor and Jones<br />

2007), as we describe next.<br />

5.3 Evaluation<br />

In the evaluation step, the researcher must evaluate if the IT<br />

artefact meets the requirements and solves the real-world<br />

problem (Gregor and Jones 2007), because “What works<br />

and doesn’t work will evolve over time based upon feedback<br />

and learning from applying the ideas and analyzing the<br />

results” (Basili 1996). This step tests if the criteria that were<br />

explored in the awareness of the problem and suggestion step<br />

are accomplished. The outcome of this step is represented by<br />

performance measures for the IT artefact (Kuechler and Vaishnavi<br />

2008a, b). Both the creation and evaluation of the IT<br />

artefact are highly intertwined. Without a problem solution,<br />

the preceding steps must be repeated, until the evaluation<br />

actually indicates a satisfactory problem solution. Thus,<br />

theory-generating DSR includes a refinement cycle of data<br />

collection, development, and creation. For traditional<br />

approaches, a solution signals the end (Aalst and Kumar<br />

2003; Umapathyetal.2008), such that the contribution to<br />

the knowledge base is the development and evaluation of the<br />

IT artefact (Gregor 2006; Hevneretal.2004). However,<br />

theory-generating DSR goes beyond that border.<br />

5.4 Theory-generation<br />

Beside the simultaneous use of GTM and DSR in the<br />

awareness and suggestion step, theory-generating DSR enables<br />

the researcher to conduct an additional theorygeneration<br />

step. In our model, this step represents an extension<br />

of Kuechler and Vaishnavi’s (2008a, b). It enables<br />

additional theoretical insights, beyond the developed IT<br />

artefact and therefore represents an intertwined process of<br />

problem solving and theorizing. In particular, these additional<br />

insights present the outcomes of this extension. It also


uncovers the potential of GTM to offer valuable research<br />

advice, e.g., techniques such as theoretical sampling and<br />

constant comparative method, to a growing area of interest<br />

in IS research that focuses on the IT artefact, as demanded by<br />

leading scholars (e.g., Hevner et al. 2004; Orlikowskiand<br />

Iacono 2001). GTM focuses on generating a behavioural<br />

understanding of the focal phenomenon and making a theoretical<br />

contribution to the knowledge base, whereby in the<br />

prior steps, DSR solved a real-world problem and developed<br />

the basis for this analysis; the IT artefact.<br />

While GTM enables DSR to offer an additional theoretical<br />

contribution to the knowledge base beyond the IT artefact<br />

(Hevner et al. 2004), DSR enables GTM to bring the IT<br />

artefact to the centre of scholarly attention (Orlikowski and<br />

Iacono 2001).<br />

5.4.1 Additional theoretical sampling<br />

Several additional steps provide new insights about the usage<br />

and performance of the IT artefact. A created IT artefact can<br />

change users’ behaviours and expectations, which would<br />

define new requirements for follow-up projects and improvements<br />

of the IT artefact. In this sense, the approach has created<br />

a reusable contribution to the knowledge base. Weedman<br />

(2008) thus describes a DSR project focusing on the evaluation<br />

and theory-generating part supported by GTM<br />

techniques.<br />

Additional theoretical sampling involves further data collection,<br />

after the creation of the IT artefact, with support<br />

from GTM. These additional data enable the researchers to<br />

explore the performance, usability, and assimilation of the<br />

IT artefact. This step increases understanding of the usage of<br />

the IT artefact and provides a means to explore changes in<br />

people’s behaviour after they use the IT artefact.<br />

Again, the data collection and analysis guide any subsequent<br />

data collection. Therefore, the understanding of the<br />

phenomena, the usage of the IT artefact, and its impact on<br />

people’s behaviour emerge over time, through iterative cycles<br />

of data collection and analysis (Glaser 1978). Over time, the<br />

researcher achieves theoretical saturation (Eisenhardt 1989).<br />

This extremely important step in theory-generating DSR<br />

bridges the gap between design and behavioural sciences<br />

(Holmström et al. 2009; Orlikowski and Iacono 2001). Some<br />

prior research used additional theoretical sampling but unfortunately<br />

only to evaluate the created IT artefact, without<br />

clarifying the connection between the approaches (e.g., Albert<br />

et al. 2004; Markusetal.2002).<br />

5.4.2 Additional slices of data<br />

Additional data may come from, though are not limited to,<br />

the application of the IT artefact in the appropriate environment,<br />

applicable knowledge from existing literature, and<br />

results from the IT artefact evaluation. The overall analysis<br />

moves up to a higher conceptual level (Levina and Vaast<br />

2005). Furthermore, DSR provides unique data that cannot<br />

be used in an ordinary GTM approach, in the form of socalled<br />

throw-away prototypes (e.g., Markus et al. 2002) and<br />

data from IT artefact testing. Not only does the usage of an<br />

IT artefact depend on its environment, but IT artefacts explicitly<br />

can change their environment and people’s behaviours<br />

and thus the requirements for future IT artefacts.<br />

However, throw-away prototypes offer only instantiations<br />

as IT artefacts. Thus, in theory-generating DSR, additional<br />

slices of data can contain more than literature, including<br />

prototypes and other innovative data.<br />

5.4.3 Categories<br />

An essential step that condenses core categories during the<br />

analysis involves the structured process of coding and analysis<br />

(Glaser 1978). Different models for coding exist, but in the<br />

Glaserian version, the researcher starts with so-called open<br />

coding, such that he or she groups indicators from the data or<br />

initial IT artefacts into concepts and then categories over time,<br />

when it becomes clear which themes are of central interest to<br />

yield the desired theoretical insights. The coding process itself<br />

evolves and changes over time to become more selective, such<br />

that prior conceptualizations and codes guide subsequent<br />

steps. To generate these categories (and their properties) from<br />

the collected data, researchers undergo several iterations of<br />

coding and analysis and employ constant comparison (Glaser<br />

1978). In this sense, constant comparison is the process of<br />

constantly comparing instances of slices of data with other<br />

instances of data to generate the categories. This comparison<br />

can also be applied to define and refine existing categories. It<br />

contributes to the development of theory by structuring the<br />

analytic properties of the data and categories (Urquhart et al.<br />

2010). For instance, Markus et al. (2002) use throw-away<br />

prototypes as sorted categories, refined over such iteration<br />

steps. Thus, researchers can reveal any clear fit of the IT<br />

artefacts with primary or secondary data, including people’s<br />

changed behaviours. The integration of extant literature into<br />

the coding and conceptualization process can enrich the process<br />

of creating and exploring categories. As a result, the<br />

researcher must not “force” additional theoretical insights<br />

onto the data. The additional data collection and analysis<br />

continues until a point of theoretical saturation. This point is<br />

reached when the researcher believes further data will not lead<br />

to any more insights.<br />

5.4.4 Additional theoretical insight<br />

Inf Syst Front<br />

Condensing the emerged categories allows for analyses of<br />

the relationships among them and thus increases insights<br />

(Glaser 1978). The end result, such as an added contribution


Inf Syst Front<br />

to the domain of study in the form of grounded theory about<br />

the developed IT artefact, expands the knowledge base (e.g.,<br />

Weedman (2008) contributes to social science theory). After<br />

identifying and defining the core categories, researchers<br />

must assess the relationships among them to generate additional<br />

theoretical insights. An important step to achieve the<br />

final theoretical contribution of this study entails extensive<br />

comparisons across the generated insights and prior published<br />

work in the same domain. The emergent contribution<br />

and its theoretical insights then can be integrated into<br />

follow-up DSR projects as new requirements to consider.<br />

5.5 Conclusion<br />

To encourage a cumulative research tradition and benefit<br />

from existing problem solutions, the results of the proposed<br />

research approach should be used in further projects or<br />

research cycles, integrated as an existing knowledge base<br />

and slice of data in initial iterations.<br />

In summary, theory-generating DSR integrates simultaneously<br />

techniques from DSR and GTM to construct an IT<br />

artefact and undertake an evaluation through conceptualizations<br />

and theory building (Hevner et al. 2004; Winter 2008).<br />

Prior DSR research has contributed to the knowledge base<br />

with designs and actions affiliated with the IT artefact itself<br />

(Gregor 2006). We see further potential derived from the<br />

design and action processes. Our intertwined theorygenerating<br />

approach provides a process model for such<br />

research, though the sequence of the procedure naturally<br />

must be defined by the researcher and his or her intentions.<br />

That is, theory-generating DSR is highly dynamic and<br />

should lead to a mutable IT artefact that can be easily<br />

adapted to changing environments (Gregor and Jones 2007).<br />

6 Evaluation of our approach to theory-generating<br />

design science research<br />

To evaluate the developed theory-generating DSR approach,<br />

we present first a theoretical evaluation in which we describe<br />

how the developed model fits to the IS design theory<br />

components of Gregor and Jones (2007) and second, illustrate<br />

with the help of an exemplary case of the extant DSR<br />

literature how some of the theory-generating DSR elements<br />

can be applied.<br />

6.1 Theoretical evaluation<br />

In Table 2 we provide a comparison of the proposed steps<br />

and activities against the eight IS design theory components<br />

provided by Gregor and Jones (2007). Beyond descriptions<br />

of our developed theory-generating DSR approach, we offer<br />

information about the activity in and legitimation for each<br />

step. Finally, we illustrate how the IS design theory components<br />

of Gregor and Jones (2007) relate to the steps of<br />

theory-generating DSR.<br />

The missing design theory component not addressed by<br />

our theory-generating DSR approach is component 4 from<br />

Gregor and Jones (2007): artefact mutability. Despite this<br />

apparent gap, we believe the overall process of generating<br />

additional theoretical insights from DSR finding leads to a<br />

more mutable IT artefact that can be adapted easily to<br />

changing environments. Furthermore, the detailed analysis<br />

of the IT artefact’s requirements and its functionality in a<br />

given or changing environment makes the IT artefact more<br />

accountable, flexible, and mutable; it is not necessarily<br />

embedded in some specific place, time, or community<br />

(Orlikowski and Iacono 2001). Therefore, though our approach<br />

does not explicitly deal with IT artefact mutability,<br />

which may be regarded as a limitation, we believe it<br />

addresses artefact mutability implicitly, through the process<br />

of iterating IT artefact cycles (problem solving, theoretical<br />

saturation) as the IT artefact goes through repeated cycles of<br />

change and reflections. However, because our theorygenerating<br />

DSR approach aims to achieve new insights,<br />

testing artefact mutability with an upfront chosen kernel<br />

theory admittedly is not possible.<br />

6.2 Illustration of our model through an exemplary case<br />

Weedman (2008) analyzed a design project from a behavioural<br />

and social science perspective. This design project<br />

was about the implementation of ‘Sequoia 2000’ which<br />

represents an interactive information system based on the<br />

data-handling needs in global change science. Earth scientists<br />

need to analyze a huge mass of data of different types,<br />

e.g., satellite data, text, raster, and vector with Sequoia 2000<br />

as the platform that allows to analyze and share the data<br />

among them. The development of Sequoia 2000 can be used<br />

to illustrate several steps of our model even though they are<br />

not explicitly followed by Weedman (2008).<br />

Weedman (2008) entered the field and positioned their<br />

research among others on the design theories of Gregor and<br />

Jones (2007). She conducted an extensive literature review<br />

on DSR and design theories to be informed (awareness) of<br />

the problem. However, their problem was grounded in the<br />

need for a collaboration and analysis tool for earth scientists.<br />

The requirements for the huge data-handling problem were<br />

officially stated as: vastly increased storage, much faster<br />

networking, visualization techniques for modelling, and a<br />

new database management system to replace the existing<br />

flat file system. The expectation towards the system was that<br />

it would change the way how global change researchers<br />

work, for example by allowing much more data transfer<br />

than in the past based on increased storage and network<br />

speed. However, initial theoretical sampling, such as


Table 2 Eight components of IS design theory in different DSR approaches<br />

Step Derived from Activity Reason Supported Design Theory<br />

Components<br />

Awareness of problem<br />

and suggestion<br />

DSR (Kuechler and<br />

Vaishnavi 2008a, b;<br />

Peffers et al. 2008)<br />

Theoretical sampling GTM (Glaser 1978) Systematic collection and<br />

analysis of data<br />

Slices of data GTM (Fernandez<br />

2004; Levina and<br />

Vaast 2005)<br />

Tentative design and<br />

requirements<br />

DSR (e.g., Hevner et<br />

al. 2004)<br />

Development DSR (Hevner et al.<br />

2004; Kuechler and<br />

Vaishnavi 2008a, b)<br />

Evaluation DSR (Basili 1996;<br />

Kuechler and<br />

Vaishnavi 2008a, b)<br />

Additional theoretical<br />

sampling<br />

Additional slices of<br />

data<br />

interviews with the prospective users, pointed out that their<br />

understanding of the problem was somewhat different. They<br />

thought that it would only provide them with more advanced<br />

hard- and software, without however triggering a fundamental<br />

change in earth science. For instance, a programmer working<br />

for earth scientists pointed out that he does not understand that<br />

this project is beneficial for him. To meet these moving<br />

targets, a questionnaire was sent out to all participants and<br />

several interviews were conducted. Thereby, the participants<br />

and the collected data represent environmental factors as slices<br />

of data. For instance, the interviewees were asked about<br />

potential problems with Sequoia 2000. They pointed out that<br />

some problems with the handling of the technology could<br />

occur. The interviewees based these statements on experiences<br />

Definition of research lens Need for appropriate<br />

combination of design and<br />

behavioural science<br />

Identification of necessary<br />

slices of data for both<br />

DSR and behavioural<br />

understanding<br />

Shaping requirements and<br />

business needs<br />

Development of the IT<br />

artefact<br />

Provides the means to identify<br />

requirements and a basis for<br />

generating theory from an<br />

IT artefact<br />

Allows for combination of<br />

requirement identification/<br />

business needs and<br />

theoretical conceptualization<br />

Fosters the relevance of the<br />

problem solution<br />

Strengthens the practical<br />

contribution of the DSR<br />

finding<br />

Evaluation of the IT artefact Confirms whether the<br />

problem is solved<br />

GTM (Glaser 1978) Collection and analysis of<br />

additional data based on<br />

IT artefact embedded in<br />

the environment<br />

GTM (Fernandez<br />

2004; Levina and<br />

Vaast 2005)<br />

Identification of additional<br />

slices of data for<br />

behavioural understanding<br />

Categories GTM (Glaser 1978) Generation of conceptual<br />

categories (and properties)<br />

from the intertwined data<br />

Conclusion DSR (Hevner et al.<br />

2004; Kuechler and<br />

Vaishnavi 2008a, b)<br />

collection and analysis<br />

Embedding theoretical<br />

insights into DSR<br />

literature<br />

Extends the scope of the<br />

analysis beyond business<br />

needs/requirements toward<br />

a behavioural understanding<br />

based on preliminary<br />

theoretical sampling<br />

Strengthens the theoretical<br />

conceptualization process<br />

within the DSR process<br />

Forces DSR researchers to<br />

conceptualize and reach<br />

theoretical abstraction<br />

Closes the loop by stimulating<br />

further DSR research that<br />

builds on the generated<br />

theoretical findings<br />

Inf Syst Front<br />

Component 6: Underlying<br />

theories to justify and<br />

explain the design<br />

Component 2: Through a<br />

detailed analysis, the<br />

entities of interest can be<br />

derived and clearly<br />

communicated<br />

Component 1: Defines<br />

what the system is for<br />

Component 8: The<br />

physical implementation<br />

of the IT artefact<br />

Component 5: Evaluates<br />

whether the IT artefact<br />

satisfies the requirements<br />

Component 7: Helps<br />

implement additional<br />

theoretical insights<br />

about the IT artefact<br />

through additional data<br />

collection and analysis<br />

Component 3: An<br />

additional theoretical<br />

contribution of the IT<br />

artefact and its<br />

behaviours can function<br />

as a blueprint for followup<br />

projects<br />

with the programmers who emphasized that assistance in the<br />

use of the technology was not within scope. As a consequence,<br />

the development was adapted to this insight and earth<br />

scientists were included prior to the implementation. This<br />

implied that the potential user became accustomed to the<br />

technology which resembles an intertwined process of data<br />

collection and analysis to refine the tentative design and the<br />

requirements of the technology.<br />

After solving these problems, Sequoia 2000 was developed<br />

and evaluated stepwise. The first prototypes were fed<br />

with data from the earth scientists to test the performance.<br />

Several problems occurred in this process. For instance, the<br />

network went down several times or the file system broke<br />

down because of the huge mass of data. As a consequence,


Inf Syst Front<br />

additional theoretical sampling and slices of data were<br />

collected from the participants. The programmers saw the<br />

earth scientists involved in the build and evaluate step as a<br />

kind of test bed for early systems testing and evaluation.<br />

However, the earth scientists saw their role as customers not<br />

being interested in acting as test persons but rather receive a<br />

stable and ready-to-use technology that enables them to<br />

conduct their research. The understanding of the earth scientists<br />

as customers rather than as participants in the design<br />

reinforced their research goals as being more important than<br />

the design goals. Hence, Sequoia 2000 had to go through<br />

several other improvement loops because of additional<br />

deliverables that were not fixed in the initial requirements<br />

analysis.<br />

The entire Sequoia project can be seen as an example of<br />

including behavioural science aspects into the design and<br />

development of a technology. This project generated different<br />

conclusions and thereby additions to the knowledge<br />

base. In particular, Weedman (2008) developed two different<br />

types of contributions: First, she developed an explanation<br />

of how the meta-requirements of handling large or<br />

massive amounts of data in the domain of earth science<br />

could be matched with different kinds of solution components,<br />

including a cross-disciplinary metadata schema and<br />

associated chunking strategy for database management,<br />

principles of data visualization, and the principle of data<br />

reusability for running different analysis tasks. According to<br />

Baskerville and Pries-Heje (2010), this type of contribution<br />

resembles an ‘explanatory design theory’, i.e. the design<br />

product (Walls et al. 1992). However, the second and probably<br />

more important contribution of Weedman (2008) was the<br />

development of categories and relationships between them<br />

regarding the design process of solving ill-defined or wicked<br />

problems. Such a ‘design process theory’ (Baskerville and<br />

Pries-Heje 2010) is represented in the following Fig. 3.<br />

Accordingly, different categories emerged from<br />

Weedman’s (2008) analysis which are related to the core<br />

Involvement of<br />

users as partners<br />

Collaboration<br />

Reflective designeruser<br />

conversation<br />

Interdisciplinary<br />

cooperation<br />

Mutual interests<br />

and benefits<br />

Different viewpoints<br />

and knowledge<br />

Incentives/<br />

motivation<br />

Fig. 3 Derived design process theory drawn from Weedman (2008)<br />

notion of collaborative DSR projects with ‘reflective<br />

designer-user conversations’ at its centre. Adapted from<br />

Schön (1983, 1992), it represents the idea that designers<br />

and users engage in interactive conversations about both<br />

problem formulation and solving. Such a conversation<br />

stimulates a kind of reflection-in-action and shared understanding<br />

between the designers and users which<br />

leads to enhanced design outcomes. This central aspect<br />

of the collaboration is leveraged and enabled through<br />

three further concepts. First, designers and users are<br />

motivated by placing mutual interest and benefits at<br />

the centre of team composition. According to Weedman<br />

(2008), this can be further leveraged through appropriate<br />

incentive structures. Second, the concept of interdisciplinary<br />

cooperation provides different viewpoints and<br />

knowledge into the design process. Therewith, different<br />

knowledge types to be combined, including domain expert<br />

knowledge and design knowledge. Finally, users are involved<br />

as partners in the design process to stimulate collaboration<br />

with designers. In summary—drawing from Weedman (2008)<br />

—a design process theory can be derived that addresses<br />

motivational (i.e. mutual interests and benefits), cognitive (i.e.<br />

interdisciplinary knowledge), and behavioural (i.e.<br />

designer-user collaboration) aspects.<br />

7 Discussion<br />

While DSR has become an established area of research<br />

within IS, it still lacks the maturity of more accepted<br />

research approaches in terms of set of research instruments<br />

used or evaluation of theoretical contributions made. Since the<br />

seminal work of Hevner et al. (2004) there is an even stronger<br />

debate how DSR can be combined with other research<br />

methods to increase its rigor.<br />

Theory-generating DSR focuses on the creation and evaluation<br />

of theoretical abstractions from an IT artefact by combining<br />

elements of constructive research with key concepts<br />

and techniques from interpretative research, thus allowing for<br />

grounded theorizing. Our suggested approach extends the<br />

general design cycle for DSR proposed by Kuechler and<br />

Vaishnavi’s (2008a, b). Different from Holmström et al.<br />

(2009) and Sein et al. (2011) we do not include behavioural<br />

science as an additional and independent component into DSR<br />

but rather develop a process model that enables a simultaneous<br />

and highly intertwined use of both. In so doing, we can<br />

combine the development of an IT artefact that solves a class<br />

of real-world problems in a new and innovative way with<br />

rigorous theorizing, such as about how the newly developed<br />

IT artefact interacts with its environment.<br />

In particular, the awareness of the problem and the suggestion<br />

step can be enhanced through a detailed analysis of<br />

the environment and the extant knowledge base. Through


theoretical sampling, slices of data, and the constant comparison<br />

method researchers may be able to explore the<br />

problem and the phenomenon more precisely and analyze<br />

potential failures in the development in advance. A tentative<br />

design and the requirements of the IT artefact are more<br />

detailed through this refined step in theory-generating<br />

DSR. In addition, the extension of theory-generation after<br />

the evaluation of the IT artefact enables researchers to create<br />

an additional theoretical insight, e.g., from the application of<br />

the IT artefact in the appropriate environment, applicable<br />

knowledge from existing literature, or results from the IT<br />

artefact evaluation. This additional theoretical component<br />

offers another research outcome, but it also complements<br />

the underlying DSR approach and thereby meets the<br />

requirements for a IS design theory, as specified by Gregor<br />

and Jones (2007). As an exemplary case from extant literature,<br />

we used Weedman (2008) who used different behavioural<br />

science elements to analyze their developed IT<br />

artefact in more detail.<br />

From a scientific point of view, theory-generating DSR<br />

focuses on the IT artefact and its improvement, as well as<br />

offering additional theoretical insights based on the use of the<br />

IT artefact. Thus, its reusable contribution adds to existing<br />

knowledge bases. Furthermore, it designs and creates IT artefacts<br />

as means to discover new knowledge (Baskerville et al.<br />

2009), which distinguishes theory-generating DSR from, e.g.,<br />

action research, which aims to create change in an organizational<br />

setting and studies the subsequent effects (Baskerville et<br />

al. 2009). Moreover, action researchers take an action and<br />

apply extant theories within the course of the research project<br />

(Coglan and Coughlan 2002; McKay and Marshall 2001). It<br />

thus explores the research project from an internal perspective;<br />

the action researcher works with the people directly affected<br />

by the action or who have the potential to influence the action<br />

in their environment (Avison et al. 1999). In contrast, a theorygenerating<br />

DSR researcher observes phenomena and interacts<br />

only within the scientific environment. Accordingly, theorygenerating<br />

DSR provides findings about the potential<br />

improvements to an IT artefact, and the researcher simply<br />

observes.<br />

8 Conclusion<br />

We have developed and presented a conceptual approach<br />

for a complementary use of DSR and behavioural<br />

science research elements, on the basis of Kuechler<br />

and Vaishnavi’s (2008a, b) model and comparison with<br />

components necessary for IS design theories. For our initial<br />

theory development, we used interrelations to combine previously<br />

unconnected bodies of knowledge (DSR and GTM). In<br />

the course of this combination, we identified gaps in their<br />

conceptualizations and added missing components to an<br />

already existing DSR model from Kuechler and Vaishnavi’s<br />

(2008a, b). However, the proposed approach is a general<br />

process model, rather than a strict recipe, and has not yet been<br />

challenged or tested through application to actual DSR<br />

projects.<br />

According to Hevner et al. (2004), theoretical contributions<br />

to the knowledge base represent an important and<br />

necessary part of DSR. We therefore propose the combination<br />

of DSR and GTM to unite design and behavioural<br />

aspects. In particular, we illustrate how design and behavioural<br />

research elements combine effectively in a pluralistic<br />

research design, which responds to calls to find potential<br />

research improvements (e.g., Mingers 2001).<br />

Future research in this direction should evaluate whether<br />

such a combined research method is applicable or not in<br />

DSR projects. The used behavioural science method in this<br />

paper uses elements from GTM which is only one out of<br />

many methods in the field. Hence, this paper presents a new<br />

attempt to combine design with behavioural research. Future<br />

research should evaluate further the simultaneous use of<br />

behavioural and DSR methods into a pluralistic research<br />

design.<br />

In summary, for a scientific discipline, generating contributions<br />

to its knowledge base is at least as important as<br />

solving real-world problems. We strongly believe that the<br />

proposed theory-generating DSR approach as one possible<br />

combination of design and behavioural science can support<br />

this goal by providing a process model for researchers who<br />

strive to follow and accomplish this aim in a research<br />

setting. From a theoretical perspective, this model can provide<br />

a fruitful contribution to the DSR community and<br />

expect to be very interesting to researchers focusing on<br />

DSR from a methodological point of view. From a practical<br />

perspective, managers could use the developed model for<br />

their DSR projects and thereby improve the development of<br />

prototypes in a structured way to avoid failure or repair<br />

loops.<br />

Appendix: Review of DSR Articles<br />

Inf Syst Front<br />

In the course of our literature research, we found several<br />

articles beside Weedman (2008) that offer support for various<br />

steps of our theory-generating DSR approach. Each<br />

article depicts a usage of DSR methods. Many articles<br />

present the IT artefact as an instantiation, though as Hevner<br />

et al. (2004) state: “IT artefacts can also be represented by<br />

constructs, models, methods, instantiations, or a combination<br />

thereof” (see also, March and Smith 1995). In Table 3,<br />

we list articles that discuss typical examples of DSR and that<br />

provided some basic ideas for our theory-generating DSR<br />

approach.


Inf Syst Front<br />

Table 3 Literature using DSR as a research approach<br />

Article Developed IT-Artefact Description<br />

Aalst and Kumar (2003) Interorganizational routing<br />

language based on XML<br />

Abbasi and Chen (2008) CyberGate: A design framework<br />

and system for text analysis of<br />

computer-mediated<br />

References<br />

communication<br />

Albert et al. (2004) GIST: A model of the design and<br />

management of content and<br />

interactivity of customer-centric<br />

Web sites<br />

Umapathy et al. (2008) A model for knowledge reuse for<br />

designing enterprise integration<br />

solutions<br />

Markus et al. (2002) A design theory for systems that<br />

support emergent knowledge<br />

processes<br />

Aalst, W., & Kumar, A. (2003). XML-based schema definition for<br />

support of interorganizational workflow. Inform Syst Res, 14(1),<br />

23–46.<br />

Abbasi, A., & Chen, H. (2008). CyberGate: a design framework and<br />

system for text analysis of computer-mediated communication.<br />

MIS Quarterly, 32(4), 811–837.<br />

Albert, T. C., Goes, P. B., & Gupta, A. (2004). GIST: a model for design<br />

and management of content and interactivity of customer-centric<br />

web sites. MIS Quarterly, 28(2), 161–182.<br />

Allen, D. K., Colligan, D., Finnie, A., et al. (2000). Trust, power and<br />

interorganizational information systems: the case of the electronic<br />

trading community TransLease. Inform Syst Res, 10(1), 21–40.<br />

Arazy, O., Kumar, N., & Shapira, B. (2010). A theory-driven design<br />

framework for social recommender systems. J Assoc Inform Syst<br />

Online, 11(9), 455–490.<br />

Au, Y. A. (2001). Design Science I: the role of design science in electronic<br />

commerce research. Comm Assoc Inform Syst, 7(1), 1–15.<br />

Avison, D., Lau, F., Myers, M., et al. (1999). Action research. Comm<br />

ACM, 42(1), 94–97.<br />

Basili VR (1996) The role of experimentation in software engineering:<br />

past, current, and future. Proceedings of the International<br />

Conference on Information Systems (ICIS), Ohio, USA.<br />

Baskerville, R. (2008). What design science is not. Eur J Inform Syst,<br />

17(5), 441–443.<br />

The authors describe a standard problem-oriented approach. They use<br />

DSR methods accidentally to create a new interorganizational routing<br />

language. They test this language with a prototype implementation.<br />

The article depicts the typical engineering and problem-oriented<br />

approach from which DSR is derived.<br />

The authors conduct a procedure analogous to Aalst and Kumar’s<br />

(2003) and develop CyberGate, an IT artefact for text analysis<br />

techniques.<br />

The authors describe a standard problem-oriented approach. The IT<br />

artefact is derived from existing models, methods, and additional<br />

data. The requirements are obtained from an analysis of the existing<br />

environment. The authors conduct an elaborate evaluation of the IT<br />

artefact by collecting interview and survey data among users to get a<br />

deeper insight on its usability. Moreover, they conduct a gap analysis<br />

to identify potential gaps and missing values in their created model.<br />

Unfortunately, they do not use the data they gathered to improve the<br />

IT artefact.<br />

The authors describe a standard problem-oriented approach. The<br />

requirements of the IT artefact are described by an analysis of the<br />

existing environment. The evaluation of the IT artefact involves a<br />

comparison of the old IT artefact with the new one.<br />

The authors use DSR to create their IT artefact without mentioning it.<br />

Their research offers a closely intertwined combination of DSR and<br />

behavioural science. The authors build their IT artefact on existing<br />

kernel theories and conduct interviews to ensure the usability of their<br />

IT artefact. Unfortunately, these steps were not sequential. Several<br />

iterations of building and testing the IT artefact were conducted with<br />

so-called throw-away prototypes. The article provides extremely<br />

fruitful suggestions for our theory-generating DSR approach.<br />

Baskerville, R., & Pries-Heje, J. (1999). Grounded action research: a<br />

method for understanding IT in practice. Account Manag Inform<br />

Tech, 9(1), 1–23.<br />

Baskerville, R., & Pries-Heje, J. (2010). Explanatory design theory.<br />

Business & Information Systems Engineering, 2(5), 271–282.<br />

Baskerville, R., & Stage, J. (2001). Accommodating emergent work<br />

practices: ethnographic choice of method fragments. In B.<br />

Fitzgerald, N. Russo, & J. I. DeGross (Eds.), Realigning research<br />

and practice: the social and organizational perspectives. Boston:<br />

Kluwer Academic Publishers.<br />

Baskerville R, Pries-Heje J, Venable J (2009) Soft design science<br />

methodology. Proceedings of Design Science Research in Information<br />

Systems and Technology (DESRIST), Philadelphia, USA.<br />

Benbasat, I., & Zmund, R. W. (2003). The identity crisis within the IS<br />

discipline: defining and communicating the discipline’s core<br />

properties. MIS Quarterly, 27(2), 183–194.<br />

Bryant, A., & Charmaz, K. (2007a). Grounded theory in historical perspective:<br />

an epistemological account. In A. Bryant & K. Charmaz<br />

(Eds.), The SAGE handbook of grounded theory. London: SAGE<br />

Publications.<br />

Bryant, A., & Charmaz, K. (2007b). Grounded theory research: methods<br />

and practices. In A. Bryant & K. Charmaz (Eds.), The SAGE<br />

handbook of grounded theory. London: SAGE Publications.<br />

Carlsson SA (2006) Towards an information systems design research<br />

framework: a critical realist perspective. Proceedings of Design<br />

Science Research in Information Systems and Technology<br />

(DESRIST), Claremont, USA.


Coglan, D., & Coughlan, P. (2002). Action research for operations<br />

management. Int J Oper Prod Manag, 22(2), 220–240.<br />

Dubin, R. (1978). Theory building. New York: Free Press.<br />

Eisenhardt, K. M. (1989). Building theories from case study research.<br />

Acad Manage Rev, 14(4), 532–550.<br />

Eisenhardt, K. M., & Graebner, M. E. (2007). Theory building from<br />

cases: opportunities and challenges. Acad Manage J, 50(1), 25–32.<br />

Fernandez WD (2004) The grounded theory method and case study<br />

data in IS research: issues and design. Proceedings of Information<br />

Systems Foundations Workshop: Constructing and Criticising,<br />

Canberra, Australia.<br />

Glaser, B. G. (1978). Theoretical sensitivity: advances in the methodology<br />

of grounded theory. Mill Valley: Sociology Press.<br />

Glaser, B. G. (1998). Doing grounded theory: issues and discussions.<br />

Mill Valley: Sociology Press.<br />

Glaser, B. G. (2007). Doing formal theory. London: SAGE<br />

Publications.<br />

Glaser, B. G., & Strauss, A. L. (1967). The discovery of grounded<br />

theory: strategies for qualitative research. Chicago: Aldine<br />

Publishing Company.<br />

Goldkuhl, G. (2004). Design theories in information systems—a need<br />

for multi-grounding. J Inform Tech Theor Appl, 6(2), 59–72.<br />

Gregor, S. (2002). Design theory in information systems. Aust J Inform<br />

Syst, 10(1), 14–22.<br />

Gregor, S. (2006). The nature of theory in information systems. MIS<br />

Quarterly, 30(3), 611–642.<br />

Gregor, S., & Jones, D. (2007). The anatomy of a design theory. J<br />

Assoc Inform Syst Online, 8(5), 312–335.<br />

Gregory, R. W., & Muntermann, J. (2011) Theorizing in design science<br />

research: inductive versus deductive approaches. In: Thirty second<br />

international conference on information systems, Shanghai,<br />

China.<br />

Hevner, A. R., & March, S. T. (2003). The information systems<br />

research cycle. IT Systems Perspective, 36(11), 111–113.<br />

Hevner, A., March, S., Park, J., et al. (2004). Design science in<br />

information systems research. MIS Quarterly, 28(1), 75–105.<br />

Holmström, J. B., Ketokivi, M., & Hameri, A. P. (2009). Bridging<br />

practice and theory: a design science approach. Decision Sci, 40<br />

(1), 65–87.<br />

Ketokivi, M., & Mantere, S. (2010). Two strategies for inductive<br />

reasoning in organizational research. Acad Manage Rev, 35(2),<br />

315–333.<br />

Kuechler, B., & Vaishnavi, V. (2008a). On theory development in<br />

design science research: anatomy of a research project. Eur J<br />

Inform Syst, 17(5), 489–504.<br />

Kuechler B, Vaishnavi V (2008b) Theory development in design<br />

science research: anatomy of a research project. In Proceedings<br />

of Design Science Research in Information Systems and Technology<br />

(DESRIST), Atlanta, GA.<br />

Lee JS, Pries-Heje J, Baskerville R (2011) Theorizing in design science<br />

research. 6th International Conference on Design Science Research<br />

in Information Systems and Technology, Milwaukee, WI.<br />

Levina, N., & Vaast, E. (2005). The emergence of boundary spanning<br />

competence in practice: implications for implementation and use<br />

of information systems. MIS Quarterly, 29(2), 335–363.<br />

Madill, A., Jordan, A., & Shirley, C. (2000). Objectivity and reliability<br />

in qualitative analysis: realist, contextualist and radical constructionist<br />

epistemologies. Br J Psychol, 91(1), 1–20.<br />

March, S. T., & Smith, G. (1995). Design and natural science research<br />

on information technology. Decis Support Syst, 15(4), 251–266.<br />

Markus, M. L., Majchrzak, A., & Gasser, L. (2002). A design theory<br />

for systems that support emergent knowledge processes. MIS<br />

Quarterly, 26(3), 179–212.<br />

Inf Syst Front<br />

McKay, J., & Marshall, P. (2001). The dual imperatives of action<br />

research. Inform Tech People, 14(1), 46–59.<br />

McKay J, Marshall P (2005) A review of design science in information<br />

systems. Proceedings of the Australasian Conference on Information<br />

Systems (ACIS), Sydney, Australia.<br />

Mingers, J. (2001). Combining IS research methods: towards a pluralist<br />

methodology. Inform Syst Res, 12(3), 240–259.<br />

Nagel, E. (1961). The structure of science. New York: Harcourt, Brace<br />

& World.<br />

Nunamaker, J. F., Chen, M., & Purdin, T. D. M. (1991). Systems<br />

development in information systems research. J Manag Inform<br />

Syst, 7(3), 89–106.<br />

Orlikowski, W. J., & Iacono, C. S. (2001). Research commentary:<br />

desperately seeking the ‘IT’ in IT research—a call to theorizing<br />

the IT artifact. Inform Syst Res, 12(2), 121–134.<br />

Peffers, K., Tuunanen, T., Rothenberger, M. A., et al. (2008). A design<br />

science research methodology for information systems research. J<br />

Manag Inform Syst, 24(3), 45–77.<br />

Pries-Heje, J., & Baskerville, R. (2008). The design theory nexus. MIS<br />

Quarterly, 32(4), 731–755.<br />

Schön, D. A. (1983). The reflective practitioner: how professionals<br />

think in action. New York City: Basic Books.<br />

Schön, D. A. (1992). Designing as reflective conversation with the<br />

materials of a design situation. Knowl Base Syst, 5(1), 3–14.<br />

Sein, M. K., Henfridsson, O., Puaro, S., et al. (2011). Action design<br />

research. MIS Quarterly, 35(1), 37–56.<br />

Simon, H. A. (1969). The science of the artificial. Cambridge: MIT<br />

Press.<br />

Strauss, A., & Corbin, J. (1990). Basics of qualitative research: grounded<br />

theory procedures and techniques. London: Sage Publications.<br />

Suddaby, R. (2006). From the editors: what grounded theory is not.<br />

Acad Manage J, 49(4), 633–642.<br />

Udo K (2005) ‘Emergence’ vs. ‘Forcing’ of empirical data? A crucial<br />

problem of ‘grounded theory’ reconsidered. Forum Qualitative<br />

Social Research, 6(2), Art. 27.<br />

Umapathy, K., Purao, S., & Barton, R. R. (2008). Designing enterprise<br />

integration solutions: effectively. Eur J Inform Syst, 17(5), 518–<br />

527.<br />

Urquhart, C. (2007). The evolving nature of grounded theory method:<br />

the case of the information systems discipline. In A. Bryant & K.<br />

Charmaz (Eds.), The SAGE handbook of grounded theory. London:<br />

SAGE Publications.<br />

Urquhart, C., Lehmann, H., & Myers, M. D. (2010). Putting the<br />

‘theory’ back into grounded theory: guidelines for grounded theory<br />

studies in information systems. Inform Syst J, 20(4), 357–381.<br />

Vaishnavi, V., & Kuechler, W. (2008). Design science research methods<br />

and patterns—innovating information and communication<br />

technology. Boca Reton: Auerbach Publications Taylor & Francis<br />

Group.<br />

Walls, J. G., Widmeyer, G. R., & El Sawy, O. A. (1992). Building an<br />

information system design theory for vigilant EIS. Inform Syst<br />

Res, 3(1), 36–59.<br />

Weber, S., Beck, R., & Gregory, R. W. (2012) Combining design<br />

science and design research perspectives - Findings of three<br />

prototyping projects. In: 45 Hawaii international conference on<br />

system sciences, Wailea, Maui, HI.<br />

Weedman, J. (2008). Client as designer in collaborative design science<br />

research projects: what does social science design theory tell us?<br />

Eur J Inform Syst, 17(5), 476–488.<br />

Winter, R. (2008). Design science research in Europe. Eur J Inform<br />

Syst, 17(5), 470–475.<br />

Yadav, M. S. (2010). The decline of conceptual articles and<br />

implications for knowledge development. JMarket,74(1), 1–19.


Inf Syst Front<br />

<strong>Prof</strong>. <strong>Dr</strong>. Roman Beck is the E-<strong>Finance</strong> and Services Science<br />

Chair at Goethe University, Frankfurt, Germany. His research<br />

focuses on the creation, sourcing, and management of knowledge<br />

intensive IT services and has been published in journals such as<br />

Communications of AIS, Electronic Markets, IEEE Transactions<br />

on Software Engineering, Information Systems Frontiers, Information<br />

Technology & People, and others.<br />

Sven Weber is a research assistant and doctoral student at the<br />

E-<strong>Finance</strong> <strong>Lab</strong> at Goethe University, Frankfurt, Germany. His research<br />

focuses on the combination of design science and behavioural science<br />

research and the management of communication systems. His research<br />

has been published at several IS conferences (e.g., AMCIS, DESRIST,<br />

HICSS).<br />

<strong>Dr</strong>. Robert Wayne Gregory is an Assistant <strong>Prof</strong>essor at the Chair<br />

of Electronic <strong>Finance</strong> and Digital Markets at the University of<br />

Göttingen. He holds a PhD in Information Systems from University<br />

of Frankfurt and has a focus on both behavioural and design<br />

science-oriented research. Robert has been involved in a wide<br />

range of research projects that have been conducted in collaboration with<br />

industry, especially with financial service providers and banks. His<br />

research has been published in outlets such as Information Technology<br />

& People and ICIS.


BISE – RESEARCH PAPER<br />

The Interplay of Project Control<br />

and Interorganizational Learning: Mitigating Effects<br />

on Cultural Differences in Global, Multisource ISD<br />

Outsourcing Projects<br />

The study examines how to mitigate the cultural differences inherent in global, multisource,<br />

information systems development outsourcing projects. Its main finding is that the<br />

influence of informal control and interorganizational learning on formal control does not<br />

remain constant. Rather, it changes over time, from providing operational information to<br />

reducing formal management efforts. In turn, transparency created through formal<br />

management mechanisms provides room for effective informal control mechanisms and<br />

interorganizational learning. This interplay supports the mitigation of cultural differences<br />

through the harmonization of work-related values and practices.<br />

DOI 10.1007/s12599-012-0217-5<br />

The Authors<br />

<strong>Prof</strong>. <strong>Dr</strong>. Roman Beck<br />

E-<strong>Finance</strong> <strong>Lab</strong> & Institute of<br />

Information Systems<br />

J.W. Goethe University<br />

House of <strong>Finance</strong> Grueneburgplatz 1<br />

60323 Frankfurt am Main<br />

Germany<br />

rbeck@wiwi.uni-frankfurt.de<br />

url: http://www.servicesscience.de<br />

<strong>Dr</strong>. Katharina Schott (�)<br />

Institute of Information Systems<br />

J.W. Goethe University<br />

House of <strong>Finance</strong> Grueneburgplatz 1<br />

60323 Frankfurt am Main<br />

Germany<br />

katharina.schott@googlemail.com<br />

Received: 2011-06-26<br />

Accepted: 2012-02-14<br />

Accepted after two revisions by <strong>Prof</strong>.<br />

Leidner.<br />

Business & Information Systems Engineering<br />

This article is also available in German<br />

in print and via http://www.<br />

wirtschaftsinformatik.de: Schott K,<br />

Beck R (2012) Das Zusammenspiel<br />

von Projektsteuerung und interorganisationalem<br />

Lernen und<br />

dessen Effekt auf kulturelle Unterschiede<br />

in globalen ISE-Outsourcing-<br />

Projekten mit mehreren Dienstleistern.<br />

WIRTSCHAFTSINFORMATIK. doi:<br />

10.1007/s11576-012-0324-4.<br />

© Gabler Verlag 2012<br />

1Introduction<br />

Worldwide, companies benefit from sophisticated<br />

sourcing strategies that rely<br />

on both near- and offshore destinations.<br />

Unlike offshore outsourcing, nearshore<br />

outsourcing aims to mitigate offshorespecific<br />

challenges, such as significant<br />

time zone differences or language barriers,<br />

as well as to exploit nearshorespecific<br />

advantages, such as closer interactions<br />

through geographic proximity<br />

(e.g., Meyer and Stobbe 2007). Internationally<br />

operating vendors thus increasingly<br />

take advantage of hybrid global delivery<br />

models and organize service delivery<br />

across off-, near-, and onshore locations<br />

(Willcocks et al. 2007). Yet organizing<br />

smooth global service delivery<br />

remains challenging, especially in constellations<br />

in which client companies<br />

deal with multiple, globally distributed<br />

vendors on a single project. For example,<br />

in global, multisource projects<br />

for information systems development<br />

(ISD), both national cultures and multiple<br />

organizational cultures must converge.<br />

The perceived cultural distance between<br />

client and vendor thus increases,<br />

which requires more integrated management<br />

approaches to address greater demand<br />

for communication and coordination<br />

(Carmel and Agarwal 2001; Hildenbrand<br />

et al. 2007).<br />

Prior research in the global ISD<br />

and outsourcing domains spans multiple<br />

streams and thus focuses on various<br />

management aspects. Therefore, a<br />

large part of empirical research in global<br />

IS outsourcing analyzes cultural differences<br />

on the national, organizational,<br />

and individual level and how to deal<br />

with these differences from a project<br />

management point of view (e.g., David<br />

et al. 2008; Iacovou and Nakatsu 2008;<br />

Winkler et al. 2007). In this context, it<br />

has been shown that in particular control<br />

mechanisms as well as interorganizational<br />

learning contribute to the mitigation<br />

of cultural differences (Gregory<br />

2010b).<br />

Literature in the global ISD and outsourcing<br />

domains covers numerous studies<br />

considering control issues, such as the<br />

role of contracts (e.g., Gopal and Sivaramakrishnan<br />

2008; LacityandWillcocks<br />

1998) or other formal control mechanisms<br />

(e.g., Tiwana 2008), as well as


BISE – RESEARCH PAPER<br />

the input of informal control mechanisms<br />

(Holmström Olsson et al. 2008).<br />

Also, there are several studies focusing<br />

on interorganizational learning, mainly<br />

in relation to cross-cultural (Nicholson<br />

and Sahay 2001; Vlaar et al. 2008; Walsham<br />

2002) and knowledge (e.g., Kotlarsky<br />

et al. 2008; Leonardi and Bailey<br />

2008; Nicholson and Sahay 2004) issues.<br />

However, while these studies cover<br />

control mechanisms and interorganizational<br />

learning separately, the existing literature<br />

to the best of our knowledge<br />

has not investigated in more detail on<br />

the relationship between control mechanisms<br />

and interorganizational learning<br />

against the background of cultural differences<br />

(Gregory 2010b), in particular<br />

in an analysis that features both the<br />

client’s and multiple vendors’ perspectives.<br />

Seeking to address this gap, we adopt<br />

an exploratory research design and analyze<br />

the interplay of formal and informal<br />

control with interorganizational learning<br />

in global, multisource, ISD outsourcing<br />

projects. Accordingly, we seek to answer<br />

the following research question:<br />

How do formal and informal control<br />

mechanisms and interorganizational<br />

learning interact and contribute<br />

to the mitigation of cultural<br />

differences in global, multisource,<br />

ISD outsourcing projects?<br />

To answer this question, we apply an exploratory<br />

case study design with a global,<br />

multisource, ISD outsourcing project initiated<br />

by a large German financial institute<br />

as the research object. The objective<br />

of this project was to reengineer the<br />

financial institute’s online banking systems,<br />

and the project team included approximately<br />

100 people from five different<br />

organizations (client and four vendors),<br />

distributed across Germany, Spain,<br />

Brazil, and India.<br />

We begin our account of this project<br />

and the findings based on it by presenting<br />

theoretical foundations for our understanding<br />

of formal and informal control,<br />

as well as interorganizational learning.<br />

After we describe the case study, the underlying<br />

research methodology, and the<br />

analysis results, we discuss the findings in<br />

the light of previous literature. The final<br />

section summarizes key findings and provides<br />

some implications for research and<br />

practice.<br />

2 Theoretical Background:<br />

Control Dynamics and<br />

Interorganizational Learning<br />

Prior studies on domestic IS outsourcing<br />

emphasize the general importance<br />

of social capabilities, which not only facilitate<br />

intra-organizational cooperation<br />

but also foster mutual trust and performance<br />

among client organizations and<br />

external IS vendors (Dibbern et al. 2003).<br />

However, a recent review of IS outsourcing<br />

practices reveals that global IS outsourcing<br />

projects actually cope with even<br />

more challenges (Lacity et al. 2009). In<br />

the global context of ISD outsourcing,<br />

the need for social competency is significant<br />

because individual participants<br />

must cope with the global sourcingspecific<br />

distance between the client and<br />

vendor. This so called client-vendor distance<br />

encompasses not only geographic,<br />

economic, and political distance but also<br />

and in particular cultural differences<br />

(Vogt et al. 2009;Winkleretal.2007).<br />

In this regard governance can be conducted<br />

by the joint reliance on different<br />

governance or control mechanisms such<br />

as informal and formal control mechanisms<br />

for the governance of economic<br />

transaction (Adler 2001; Bradach and Eccles<br />

1989; Cannon et al. 2000). In the<br />

global IS outsourcing domain it has been<br />

shown that successful control balancing<br />

using combinations of different control<br />

modes is a promising way to manage cultural<br />

differences inherent in such projects<br />

(Gregory 2010a).<br />

Literature on control suggests two basic<br />

modes: formal and social control.<br />

Formal control involves the establishment<br />

and application of codified rules,<br />

goals, and procedures to define, monitor,<br />

and evaluate performance (Das and Teng<br />

2001). They usually involve explicit information<br />

transfers and include for example<br />

formal reporting guidelines and frequent<br />

meetings between key representatives<br />

(Inkpen and Currall 2004). Prior<br />

research in global IS outsourcing has<br />

employed a control perspective to show<br />

how the contract between the client and<br />

the vendor is enforced via formal behavior<br />

and outcome control; accordingly,<br />

the client controls the behavior and outcomes<br />

of his vendor (Choudhury and<br />

Sabherwal 2003). Social control involves<br />

trust-based mechanisms and operates via<br />

the development of shared values and<br />

norms (Das and Teng 2001). According<br />

to Das and Teng (1998), the key difference<br />

between formal and social control<br />

is that “formal control is more of a strict<br />

evaluation of performance while social<br />

control is about dealing with people.”<br />

(p. 501). Examples for social controls include<br />

socialization, training, and spontaneous<br />

interactions between representatives<br />

of the exchange partners (Das and<br />

Teng 1998). In the global IS outsourcing<br />

literature, social control has most often<br />

been conceptualized as informal control<br />

(Choudhury and Sabherwal 2003; Kirsch<br />

1996).<br />

An expanding body of research notes<br />

control dynamics and examines how<br />

and why control modes change across<br />

project phases. For example, Choudhury<br />

and Sabherwal (2003) analyzetheevolution<br />

of control portfolios in information<br />

systems development (ISD) outsourcing<br />

projects and reveal several factors that influence<br />

the choice and evolution of the<br />

control mechanisms. To extend such research,<br />

Kirsch (2004) describes how control<br />

modes change during three project<br />

phases of internal global IS deployment<br />

projects. Changes in the control mode<br />

appear triggered by factors from three<br />

categories: project context, stakeholder<br />

context, and global context.<br />

Besides control, also interorganizational<br />

learning has been shown to contribute<br />

to the mitigation of cultural differences<br />

in global IS outsourcing projects.<br />

On the one hand, interorganizational<br />

learning comprises the accumulation of<br />

relevant business, functional, and clientspecific<br />

knowledge, as vendors must accumulate<br />

business knowledge about their<br />

client’s application domain, functional<br />

knowledge about the client’s IT infrastructure<br />

and systems, and specific recognition<br />

of its functional requirements and<br />

processes (Dibbern et al. 2008). On the<br />

other hand, interorganizational learning<br />

pertains the important topic of knowledge<br />

transfer in global IS outsourcing<br />

projects. The existing literature repeatedly<br />

demonstrates positive effects of successful<br />

knowledge transfer (e.g., Kotlarsky<br />

and Oshri 2005; Nicholson and<br />

Sahay 2004; Oshrietal.2007; Rottman<br />

2008), and many researchers cite a lack<br />

of knowledge transfer as a major drawback<br />

for global ISD outsourcing projects<br />

(e.g., David et al. 2008; Dibbernetal.<br />

2008; Gupta and Raval 1999;Kliem2004;<br />

Leonardi and Bailey 2008). And third,<br />

interorganizational learning in a global<br />

context also refers to cross-cultural issues<br />

and the development of cultural intelligence,<br />

defined as “a form of firmlevel<br />

capability in functioning effectively<br />

Business & Information Systems Engineering


Table 1 Overview of the multisourcing portfolio<br />

Organization Area of responsibility Locations involved<br />

ARCHITECT • Definition of architectural framework<br />

• Implementation controls<br />

Two locations in Germany<br />

IMPLEMENT • Functional design Four locations in Spain<br />

• Technical design<br />

• Implementation<br />

One location in Brazil<br />

SCREEN • End-user front-end design One location in Germany<br />

TEST • Software test One location in India<br />

in culturally diverse situations” (Ang and<br />

Inkpen 2008, p. 338), as both the client<br />

and the vendor must negotiate differences<br />

in their values and work practices<br />

and learn how to adapt for the project to<br />

succeed (e.g., Carmel and Agarwal 2001;<br />

Krishna et al. 2004; Levina and Vaast<br />

2008).<br />

In summary, both control mechanisms<br />

as well as interorganizational learning<br />

have been shown to individually contribute<br />

to the mitigation of cultural differences.<br />

However, there remains little<br />

understanding of the interplay of formal<br />

and informal control mechanisms<br />

with interorganizational learning or how<br />

the interaction might contribute to mitigate<br />

cultural differences in global, multisource,<br />

ISD outsourcing projects. Furthermore,<br />

most of the preceding studies<br />

address dyadic client-vendor relationships,<br />

whereas the specific challenges of<br />

multisource projects remain unconsidered.<br />

3 Research Methods<br />

As outlined, we lack sufficient knowledge<br />

about the interplay of control and<br />

learning in global, multisource, IS outsourcing<br />

projects. Therefore, to increase<br />

this understanding, as well as to reveal<br />

the interactive effects on cultural differences,<br />

the present qualitative research<br />

features an in-depth, exploratory, singlecase<br />

study (Stebbins 2001;Yin2003). The<br />

subsequent sections describe the underlying<br />

case (3.1) as well as the procedures<br />

used for data collection (3.2) and analysis<br />

(3.3). A chronology of the overall research<br />

process is presented at the end of<br />

this chapter.<br />

3.1 Case Description<br />

The primary unit of analysis was a global<br />

ISD outsourcing project to reengineer a<br />

Business & Information Systems Engineering<br />

financial institution’s online banking system.<br />

The project was initiated because the<br />

old system’s technology required a high<br />

degree of costly expertise, and its maintenance<br />

was set to expire soon. Thus,<br />

the financial institution (BANK) decided<br />

to migrate its system to a new technology.<br />

To develop this new system, BANK<br />

applied a multisourcing strategy to reduce<br />

the risks of dependence. Therefore,<br />

it included four vendors in the project,<br />

as summarized in Table 1. The project<br />

started in October 2008 and finished<br />

in December 2009, successfully and on<br />

time.<br />

ARCHITECT, a German boutique consulting<br />

firm, designed the architectural<br />

framework of the new online banking<br />

system. IMPLEMENT was a leading international<br />

IT vendor for the financial<br />

services sector and operated and maintained<br />

BANK’s old online banking system.<br />

During the reengineering project,<br />

IMPLEMENT had the responsibility to<br />

create the functional and technical design<br />

documents and implement the new<br />

online banking system. To provide these<br />

services at the required quality and cost<br />

levels, IMPLEMENT chose a global delivery<br />

model that involved four locations<br />

in Spain and a captive center in Brazil.<br />

Another vendor, SCREEN, specialized in<br />

web development and was based in Germany.<br />

It was responsible for the frontend<br />

screen design. Finally, a large IT vendor<br />

with international operations, TEST<br />

was responsible for the software testing,<br />

conducted in a testing facility in India.<br />

Thus, the project featured both nearshore<br />

and offshore outsourcing.<br />

Thus BANK’s sourcing strategy for this<br />

project featured not only a global context<br />

but also the involvement of multiple<br />

vendors. In addition to national cultural<br />

differences resulting from the geographically<br />

distributed setting, the various organizational<br />

cultures played major roles<br />

in the project. These national and organizational<br />

cultural differences became<br />

BISE – RESEARCH PAPER<br />

especially visible in work-related values<br />

and practices, as the following examples<br />

indicate.<br />

In particular, BANK’s project manager<br />

recognized the differences on the<br />

national level:<br />

There are also cultural aspects influencing<br />

the cooperation. Due to<br />

thetimepressure,werealizedeven<br />

more that our Spanish colleagues<br />

have another understanding of milestones<br />

and time planning. During<br />

the work day, they sometimes spend<br />

two hours for a coffee break; this is<br />

very different from Germany. With<br />

detailed progress tracking, we managed<br />

to at least communicate our<br />

expectations regarding timing and<br />

quality very clearly. This helped a<br />

lot.<br />

His colleague, a sub-project manager,<br />

also cited differences that arose when<br />

working simultaneously with project<br />

teams from Spain and India:<br />

When it comes to communication, it<br />

is important to differentiate whether<br />

you are talking with a Spanish colleague<br />

or with somebody from India.<br />

When talking to an Indian colleague,<br />

you have to specify 100 %<br />

what you expect, then you also get<br />

100 %. When talking to a Spanish<br />

colleague, you have to specify what<br />

you do not want to get what you<br />

want.<br />

Beyond national cultural differences, organizational<br />

cultural differences became<br />

visible in the newly established intervendor<br />

cooperation. All the vendors previously<br />

had been involved in projects<br />

with BANK, but they had not interacted<br />

with one another in a multivendor setting<br />

before. Against this background, different<br />

work practices came to the surface<br />

and challenged the cooperation. For<br />

example, ARCHITECT was represented<br />

by a team of five experienced experts<br />

who followed an onshore delivery model.<br />

IMPLEMENT’s 50-person implementation<br />

team spread across five geographic<br />

locations and possessed various competences<br />

and practice levels. ARCHITECT<br />

thus initially had trouble understanding<br />

the challenges that faced IMPLEMENT’s<br />

large, distributed project team, including<br />

the need to scale the work to a<br />

distributed team of developers in both<br />

Brazil and Spain. A project manager from<br />

ARCHITECT explained:


BISE – RESEARCH PAPER<br />

Table 2 Demographics of<br />

interview partners<br />

Table 3 Distribution of interviews across organizations<br />

Organization Overall work experience<br />

(average)<br />

BANK (9 informants) 15 years 7 years<br />

ARCHITECT (2 informants) 6 years 6 years<br />

IMPLEMENT (9 informants) 14 years 9 years<br />

SCREEN (2 informants) 7 years 5 years<br />

Organization Organizational level/role in the project Number of people<br />

interviewed<br />

BANK (Client) Top Management 2 2<br />

Project & Sub-Project Management 5 6<br />

Project Team 2 2<br />

ARCHITECT (Vendor) Top Management 1 1<br />

Project & Sub-Project Management 1 1<br />

Project Team 0 0<br />

IMPLEMENT (Vendor) Top Management 3 4<br />

Project & Sub-Project Management 5 6<br />

Project Team 1 1<br />

SCREEN (Vendor) Top Management 1 1<br />

Project & Sub-Project Management 1 1<br />

Project Team 0 0<br />

Total 22 25<br />

Software architecture has a quite<br />

comprehensive character; if you<br />

want to understand, you need to see<br />

the overall picture, whereas IMPLE-<br />

MENT follows a “split the task and<br />

distribute the sub-tasks to the developers”<br />

mode. The developers never<br />

see the big picture.<br />

IMPLEMENT’s managers confirmed this<br />

discrepancy in work practices and explained<br />

that their global delivery model<br />

required them to look at the system at<br />

large first, and then split up the task into<br />

preferably self-contained sub-tasks. Next<br />

they could match the sub-tasks with the<br />

competence profiles of different developers<br />

to distribute the work effectively.<br />

Therefore, the distributed teams worked<br />

mostly independently – as was crucial for<br />

the geographically distributed setup.<br />

3.2 Data Collection<br />

The data collection phase comprised<br />

both interviews (primary data) and documents<br />

generated during the course<br />

of the project (secondary data). Thus,<br />

we supplemented and triangulated our<br />

interview data with project presentations,<br />

tracking sheets, status reports, and<br />

lessons-learned documents to create a<br />

rich data set. In the course of the primary<br />

data collection (taking place in July<br />

and August 2009 as well as in November<br />

and early December 2009), we conducted<br />

25 interviews with 22 respondents<br />

at both client and vendor locations.<br />

We present the demographic information<br />

about these interview partners in<br />

Table 2.<br />

Except for one conference call with a<br />

respondent located in Brazil, all the interviews<br />

took place in person: 15 in Germany<br />

and 9 in Spain. Each interview<br />

lasted between one and two hours, such<br />

that we obtained more than 38 hours<br />

of interviews. Because BANK’s corporate<br />

policy did not permit recordings of any<br />

interviews (with either BANK’s or the<br />

vendors’ employees), we took extensive<br />

notes, ultimately producing more than<br />

130 pages of write-up notes.<br />

Time spent working<br />

for the company (average)<br />

Number of interviews<br />

conducted<br />

In terms of the distribution of interviews<br />

across organizations, differences<br />

resulted from different team sizes for the<br />

parties involved. For example, ARCHI-<br />

TECT’s project team included five experienced<br />

experts who were located onshore,<br />

whereas IMPLEMENT’s team at<br />

times consisted of more than 50 people<br />

with various competencies and practice<br />

levels who worked in geographically distributed<br />

areas, including both near- and<br />

offshore locations. Table 3 clarifies the<br />

distribution of interviewees.<br />

The interview partners represented different<br />

organizational levels in BANK and<br />

three of the four vendor companies, with<br />

various profiles and roles that implied<br />

different skill and knowledge levels. We<br />

unfortunately did not have an opportunity<br />

to interview any representatives from<br />

TEST, despite repeated inquiries. The diversity<br />

of respondents ensures that our<br />

study features various organizational perspectives,<br />

hierarchical perspectives, and<br />

professional perspectives. The interviews<br />

were conversational in nature, conducted<br />

Business & Information Systems Engineering


using an interview guideline with semistructured<br />

questions. We transcribed our<br />

field notes after each interview session<br />

and used these notes to identify appropriate<br />

questions for subsequent interviews.<br />

Thus, we refined the interview<br />

questions multiple times during the<br />

course of the data collection and analysis,<br />

especially when we realized needs<br />

for additional information to confirm<br />

emerging themes or substantiate initial<br />

findings.<br />

3.3 Data Analysis<br />

The data analysis began with coding of<br />

the interview write-up notes. We identified,<br />

named, and categorized phenomena<br />

related to our research question,<br />

through comparisons of the interviews<br />

with one another and the available secondary<br />

data. The preliminary codes included<br />

concepts such as “project setup”<br />

or “initialization phase.” After we identified<br />

main conceptual themes from the interview<br />

data, according to their high frequency,<br />

we again compared all the interview<br />

data with the conceptual themes to<br />

find additional quotes or parallel statements<br />

from other interviewees. Thus, we<br />

substantiated our findings and managed<br />

to identify the main management mechanisms<br />

and learning issues from the primary<br />

data. Near the end of the analysis<br />

process, we compared our findings with<br />

the available literature, to conceptualize<br />

the emerging themes in our data. During<br />

this comparison, we elaborated on the<br />

identified themes by developing the descriptive<br />

categories into more meaningful<br />

notions at a higher level of abstraction.<br />

Thus the research team engaged in<br />

intensive considerations throughout the<br />

processtoensurenoexistingtheorywas<br />

forced onto the data. For the coding and<br />

conceptualization process, we used AT-<br />

LAS.ti. Table 4 illustrates the chronology<br />

of the overall research process.<br />

4Findings<br />

From this exploratory case study, two key<br />

findings emerged. The first finding pertains<br />

the interplay of formal and informal<br />

controls and interorganizational learning<br />

in global, multisource, ISD outsourcing<br />

projects (Sect. 4.1), while the second<br />

finding relates to the mitigating effect of<br />

these mechanisms on cultural differences<br />

in such projects (Sect. 4.2). In the following,<br />

we explain how these findings arose,<br />

Business & Information Systems Engineering<br />

describe them, and support the findings<br />

with illustrative empirical quotes from<br />

the case study interviews.<br />

4.1 Finding 1: Interplay of Formal and<br />

Informal Controls and<br />

Interorganizational Learning<br />

We start the description of this finding<br />

with the influence of formal controls on<br />

informal control and interorganizational<br />

learning: our data suggest that implementing<br />

formal control mechanisms encourages<br />

the emergence of both informal<br />

control mechanisms and interorganizational<br />

learning processes because the formal<br />

controls create transparency about<br />

the essential project parameters.<br />

In the focal project for example, the<br />

reengineering had tight deadlines because<br />

the software support for the old<br />

system’s underlying technology was set<br />

to expire. Therefore, tight project management<br />

was of particular importance,<br />

and BANK used an in-depth work breakdown<br />

structure, or traceability matrix, as<br />

a formal control mechanism. The traceability<br />

matrix originally came from the<br />

project plan, which was compiled by all<br />

involved parties. To cope with emerging<br />

differences in quality perceptions and accuracy,<br />

the parties enhanced the work<br />

breakdown structure to reflect work tasks<br />

with a very high level of detail. According<br />

to project manager,<br />

Our clear focus was on tight control<br />

of the project’s overall progress to ensure<br />

that all involved parties were<br />

on track, working toward a successful<br />

overall service delivery. Therefore,<br />

we used a very detailed progress<br />

sheet that showed for each single<br />

component when it had to be created,<br />

by whom, and according to<br />

which quality measures. Moreover,<br />

the sheet enabled the project team to<br />

identify and handle a series of delays<br />

resulting from significant task dependencies<br />

due to the involvement of<br />

multiple vendors. As a consequence,<br />

we had to track more than 1000<br />

milestones and interfaces.<br />

This traceability matrix fulfilled its function<br />

as a formal control mechanism because<br />

it helped keep track of the overall<br />

project progress and quickly identified<br />

plan variances and the need for countermeasures.<br />

Moreover, it influenced the<br />

emergence of informal control mechanisms<br />

and interorganizational learning.<br />

Specifically, by using the in-depth<br />

BISE – RESEARCH PAPER<br />

traceability matrix, the involved parties<br />

gained detailed insights into their different<br />

areas of responsibility, the associated<br />

work portfolios, and the resulting<br />

tasks for their sub-teams; they<br />

also recognized cross-organizational interdependencies.<br />

The interviewees indicated<br />

that such transparency significantly<br />

contributed to the lack of conflict<br />

or even bargaining about responsibilities<br />

and task assignments. Rather,<br />

the project team could direct its primary<br />

focus toward establishing a multisource<br />

cooperation, in terms of both growing<br />

as a team and establishing an integrated<br />

service delivery process. A project manager<br />

from one of the vendor companies<br />

explained the absence of bargaining as<br />

follows:<br />

It is essential to fully understand the<br />

project’s objectives and its planning<br />

and to always give them top priority.<br />

This may also mean that we as a<br />

vendor have to concede at a certain<br />

point. But because we are concentrated<br />

on the benefit of the overall<br />

project goals, we accept this without<br />

discussion.<br />

Although BANK and the vendor companies<br />

had never interacted in a multisourcing<br />

setting before, they developed<br />

a joint mindset with shared norms and<br />

values, including a cross-organizational<br />

team spirit and absolute goal orientation.<br />

Thus, the formal control created<br />

room for informal controls to<br />

emerge.<br />

The emergence of the joint mindset<br />

(i.e., an informal control mechanism) is<br />

perhaps clearest in IMPLEMENT’s altered<br />

attitude. When the project started,<br />

IMPLEMENT expressed a different selfconcept<br />

than the other vendors because<br />

of its long and intense prior cooperation<br />

with BANK. This vendor was accustomed<br />

to a great deal of autonomy in its implementation<br />

activities, as well as limited<br />

control by BANK, the client.<br />

In this focal project though, the multisourcing<br />

constellation moved certain<br />

tasks that IMPLEMENT had previously<br />

performed, such as the architectural<br />

framework and software testing, to other<br />

vendors. Furthermore, IMPLEMENT’s<br />

design and implementation activities no<br />

longer were controlled by BANK but<br />

instead by another, partly competing<br />

vendor (ARCHITECT). Thus, IMPLE-<br />

MENT’s project team needed to undergo<br />

a change in mindset about not only


BISE – RESEARCH PAPER<br />

Table 4 Chronology of research process<br />

No Process step Main features Additional description Result<br />

1 Literature<br />

analysis<br />

2 Identification<br />

of research<br />

case<br />

3 Data<br />

collection<br />

4 Creation of<br />

interview<br />

notes<br />

• Identification of relevant literature<br />

streams<br />

• In-depth analysis of literature within<br />

these streams<br />

• Definition of central requirements<br />

for the phenomenon<br />

• Selection of suitable case, in<br />

cooperation with industry partner<br />

• Identification of interviewees<br />

• Interviews<br />

• Collection of secondary data<br />

• Creation of clean copies<br />

• Complement interview notes with<br />

recalled details<br />

• Addition of comments<br />

5 Data analysis • Open coding<br />

• Grouping codes into categories and<br />

identifying major conceptual themes<br />

• Refinement of concepts<br />

6 Validation of<br />

findings<br />

• Discussion of researchers’<br />

interpretations with selected interview<br />

partners<br />

the content (i.e., learning the new software<br />

architecture) but also their company’s<br />

role and responsibilities in the<br />

multiparty cooperation. IMPLEMENT’s<br />

project manager described the change,<br />

and the team’s goal orientation, as follows:<br />

For sure, we would prefer to have<br />

the responsibility for the framework,<br />

as it is part of our portfolio, but the<br />

client has decided differently..., so<br />

now we are jointly responsible for<br />

this project. As a consequence, we<br />

have to prioritize the project goals<br />

more than our goals as a service<br />

delivery company to create a winwin<br />

situation. This regularly also involves<br />

overlooking the respective organizational<br />

affiliation and thinking<br />

of an integrated project team<br />

with a joint goal.<br />

• Literature streams covered:<br />

–GlobalISsourcing<br />

– IS project management<br />

• Initial interviewees suggested by senior<br />

management; further interviewees<br />

identified by the initial interviewees<br />

• 25 in-depth interviews in Germany and<br />

Spain<br />

Intensivenotetaking<br />

• Access to documents generated during<br />

thecourseoftheproject<br />

• Descriptions of the atmosphere during<br />

the interview and emotions of interview<br />

partners<br />

• Reading of transcripts and documents<br />

and highlighting of descriptions<br />

associated with the research question<br />

• Central criterion: frequency of<br />

mentions<br />

• Repeated comparisons of concepts with<br />

interview data<br />

• In the late analysis process, comparison<br />

of concepts with relevant literature<br />

• Presentation of major findings and<br />

assessments of robustness, according to<br />

interviewees<br />

This adoption of a corporate projectoriented<br />

mindset appeared among the<br />

other vendors too, as the following<br />

quote from ARCHITECT’s senior manager<br />

confirmed:<br />

In my opinion, the core team did a<br />

very good job concerning the multiparty<br />

cooperation. I sensed a broad<br />

willingness to act informally instead<br />

of insisting upon contractually<br />

agreed details and defined responsibility<br />

areas. In my experience,<br />

this is the fundamental ingredient<br />

that in the end makes projects<br />

successful.<br />

With regard to the effect of formal controls<br />

on interorganizational learning, we<br />

turn to an example from the ramp-up<br />

phase, when the vendors practiced and<br />

harmonized their cooperation and seized<br />

on opportunities to adjust their mutual<br />

• Motivation<br />

• Research gap<br />

• Specific research question<br />

• Theoretical background<br />

• Requirements list<br />

• Specific unit of analysis<br />

for study<br />

• In total: 22 interviewees<br />

• In total: Approximately 38<br />

hours of interviews<br />

• Project tracking sheets,<br />

project presentations, status<br />

reports, lessons learned<br />

• In total: Approximately<br />

130 pages of interview notes<br />

• Initial code list<br />

• Initial concepts<br />

• Final concepts<br />

expectations and needs. By practicing<br />

and reinforcing the allocation of tasks<br />

and required interactions during a fixed<br />

period at the beginning of the project,<br />

BANK ensured that the involved parties<br />

understood and accepted their own<br />

roles and responsibilities, as well as those<br />

of the other parties. All the involved<br />

parties perceived this effort as valuable.<br />

For example, IMPLEMENT’s project lead<br />

explained:<br />

We were very interested in getting<br />

this project up and running fast<br />

in order to prove our capabilities<br />

and establish a trust-based relationship<br />

with the other vendors. [...]<br />

Our objective was to first understand<br />

their organizational cultures<br />

and then partly adapt ourselves to<br />

them in order to foster a smooth<br />

integrated service delivery.<br />

Business & Information Systems Engineering


In global, multisource, ISD outsourcing<br />

projects, the responsibility for service delivery<br />

is distributed, so cooperation between<br />

vendors must be established to<br />

ensure that individual service deliveries<br />

from the different areas of responsibility<br />

intertwine and lead to a successful overall<br />

service delivery. A member of BANK’s<br />

project team summarized this essential<br />

phase:<br />

In the course of this phase, we practiced<br />

and evaluated, based on selected<br />

business transactions, how the<br />

overall service delivery had been set<br />

up and how the performance was<br />

in terms of process and outcome<br />

quality.<br />

The initialization of the multiparty cooperation<br />

also revealed performance deficits<br />

that could be attributed to national<br />

and organizational cultural differences.<br />

Such differences became manifest as divergent<br />

working modes, including values<br />

(e.g., sense of quality) and practices<br />

(e.g., knowledge transfer approaches).<br />

The project team actively addressed any<br />

conflicts or errors and used them as<br />

learning tools to improve subsequent<br />

interactions.<br />

We now describe the influence of informal<br />

controls and interorganizational<br />

learning on formal controls. It emerged<br />

from our data that in fact this influence<br />

is two-part. In the short-term, both<br />

informal control and learning mechanisms<br />

generated valuable operational information<br />

that enabled the parties to adjust<br />

their formal control mechanisms on<br />

a granular level. Then in the mid- to<br />

long-term, the informal control mechanisms<br />

and interorganizational learning<br />

processes contributed to lessen the need<br />

for formal controls. That is, our data<br />

indicated that the impact changes over<br />

time. In the following, we provide two<br />

examples to describe this effect.<br />

During the ramp-up phase, as we<br />

noted previously, the vendors practiced<br />

and adjusted their cooperation to ensure<br />

a smoothly integrated service delivery.<br />

The parties therefore identified functional,<br />

process, and technological issues<br />

and noted performance deficits, which<br />

enabled them to deduce important operational<br />

information and then sharpen<br />

their project tracking (i.e., short-term effect<br />

on formal control). A project team<br />

member explained:<br />

From my point of view, the rampup<br />

phase at the beginning of the<br />

Business & Information Systems Engineering<br />

project was very important and reasonable.<br />

As we had to deliver very<br />

early, we had to deal with problems<br />

very early as well. As a consequence,<br />

the impact of these problems<br />

could be minimized through<br />

rescheduling, and further mitigation<br />

measures could be initiated.<br />

At the beginning of the project, BANK<br />

strongly encouraged cooperation among<br />

the vendors to reinforce their roles and<br />

responsibilities and foster a stable working<br />

mode. However, this coordination<br />

and control effort gradually decreased,<br />

replaced by self-organizing mechanisms<br />

within the increasingly well-established<br />

multivendor cooperation (i.e., mid-tolong<br />

term effect on formal control).<br />

Thus the reduction of formal management<br />

effort resulted from interorganizational<br />

learning processes, as described by<br />

a project manager from IMPLEMENT:<br />

At the beginning, [BANK] arranged<br />

formal meetings managed by<br />

BANK, but later, there was more<br />

and more direct interaction between<br />

the vendor companies. BANK was<br />

not the driver but was always informed<br />

to sustain transparency.<br />

4.2 Finding 2: Mitigating Effects on<br />

Cultural Differences<br />

We now describe the mitigating effect<br />

of the interplay between control mechanisms<br />

and interorganizational learning<br />

on cultural differences in global, multisource,<br />

ISD outsourcing projects. It<br />

emerged from our data that these mechanisms<br />

can help partners overcome national<br />

and organizational cultural differences<br />

by harmonizing their varied workrelated<br />

values and practices. That is, our<br />

data indicated that the integrated use<br />

of formal and informal control and interorganizational<br />

learning dominate or<br />

overrule patterns of behavior which are<br />

rooted in national or organizational cultural<br />

differences; thus, differences in national<br />

and organizational cultures become<br />

less salient and the occurrence of<br />

culture-induced conflict declines. In the<br />

following, we describe this finding in<br />

more detail.<br />

In the focal project, the interplay of formal<br />

and informal controls, together with<br />

interorganizational learning, helped mitigate<br />

the risks associated with national<br />

and organizational cultural differences.<br />

By creating transparency about each individual<br />

deliverable (i.e., through the use<br />

BISE – RESEARCH PAPER<br />

of the formal control mechanism “traceability<br />

matrix”), the parties assimilated<br />

their divergent assessments of quality and<br />

accuracy, at least to some extent. One<br />

project manager for the vendor stated:<br />

The traceability matrix helped to<br />

mitigate the risks caused by cultural<br />

differences by defining clear roles<br />

and responsibilities, supporting the<br />

identification of interdependencies,<br />

and specifying our joint deliverables.<br />

In the following illustrative quote, another<br />

Spanish project team member explained<br />

the above-mentioned cultural<br />

differences between Germany and Spain<br />

from a Spanish perspective:<br />

When problems occur, we would expect<br />

the customer to be near you,<br />

helping you, offering his support to<br />

jointly solve the problem regardless<br />

of the timeline.<br />

However, in the course of the cooperation,<br />

it turned out that<br />

When you deliver to Germans, you<br />

have to deliver absolutely on time<br />

and bug-free. The quality expectations<br />

are high. They [German colleagues<br />

from BANK and other vendors]<br />

mainly insist on on-time delivery<br />

with defined quality; they stick<br />

more to their plan. Thus, it was very<br />

helpful that there was this detailed<br />

tracking tool, as we could see the status<br />

and our forthcoming tasks at any<br />

time.<br />

In parallel, divergent norms and values<br />

were being renegotiated and consolidated<br />

through the use of informal<br />

control and interorganizational learning<br />

processes. For example, while the<br />

development of a corporate, projectoriented<br />

mindset (i.e., informal management<br />

mechanism) helped mitigate organizational<br />

cultural differences by establishing<br />

a project culture, driven by team<br />

spirit and goal orientation, instances of<br />

interorganizational learning enabled the<br />

project team to cope with emerging differences<br />

in work-related practices. As one<br />

of our senior-level interview partners<br />

commented:<br />

From my perspective, the main reason<br />

for the success of this project<br />

is the fact that the project members<br />

from our organization and the<br />

involved vendor organizations were<br />

always willing to pursue the goals<br />

of the project in a very collaborative<br />

way.


BISE – RESEARCH PAPER<br />

Table 5 Elements of conceptual framework<br />

Underlying reasons Methods used Main results achieved<br />

• Multi-party cooperation<br />

(from both, a national and<br />

organizational perspective)<br />

• Differences in work<br />

practices<br />

• Multi-party cooperation<br />

(from both, a national and<br />

organizational perspective)<br />

• Differences in work-related<br />

values<br />

• Multi-party cooperation<br />

(from both, a national and<br />

organizational perspective)<br />

• Differences in work<br />

practices and values<br />

Categories Concepts<br />

• Formal<br />

control<br />

• Informal<br />

control<br />

• Interorganizational<br />

learning<br />

A project manager from the client organization<br />

explained in greater detail how<br />

this worked:<br />

We talked about different perceptions,<br />

divergent understandings, issues<br />

resulting from cultural differences<br />

and different communication<br />

styles, and reflected jointly on the<br />

positive and negative episodes of<br />

the cooperation between the vendors<br />

within our project, constantly<br />

aiming at further improving this<br />

cooperation.<br />

In summary, the interplay of formal and<br />

informal control and interorganizational<br />

learning enabled the project team to<br />

harmonize their divergent work-related<br />

values, norms, practices, and expectations,<br />

which resulted from national and<br />

organizational cultural differences.<br />

The findings described above are summarized<br />

in Table 5.<br />

5 Conclusion and Research<br />

Implications<br />

With the goal of increasing our understanding<br />

of ways to mitigate cultural<br />

differences in global, multisource, ISD<br />

outsourcing projects, we apply an exploratory<br />

single-case study design. In<br />

turn, we can detail how formal and informal<br />

control mechanisms and interorganizational<br />

learning interact; furthermore,<br />

our data show that these interactions<br />

help mitigating cultural differences<br />

• Traceability matrix (used as example in text)<br />

• Joint, template-based status reporting (not<br />

described in text)<br />

CMMI review process (not described in text)<br />

• Development of joint, goal-oriented<br />

mindset (used as example in text)<br />

• Socialization activities, e.g., joint dinner<br />

(not described in text)<br />

• Stimulation of teammate interaction (not<br />

described in text)<br />

• Initialization phase (used as example in<br />

text)<br />

• Use of joint structurization tools such as<br />

issue log, wiki, etc. (not described in text)<br />

in such projects. To achieve this benefit<br />

though, the implementation of formal<br />

controls appears essential because it prepares<br />

the project for the emergence of informal<br />

controls and interorganizational<br />

learning. Project partners also should anticipate<br />

changes in the effects of informal<br />

control and interorganizational learning<br />

over time, shifting from feedback and<br />

information that support the design of<br />

formal controls (short term) to an actually<br />

reduced need for and use of formal<br />

controls (mid- to long term). Together,<br />

these mechanisms can help partners<br />

overcome national and organizational<br />

cultural differences by harmonizing<br />

their varied work-related values and<br />

practices.<br />

With these findings, this study contributes<br />

to global IS outsourcing literature<br />

and provides a clearer understanding<br />

of ways to deal with cultural differences<br />

in global, multisource, ISD outsourcing<br />

projects. Furthermore, we contribute<br />

to research into control dynamics<br />

in global IS projects: this study illustrates<br />

that changes in control modes across<br />

project phases can be triggered by external<br />

factors (e.g., project context, stakeholder<br />

context, and global context as revealed<br />

by Kirsch 2004), but also through<br />

the interplay of control modes within a<br />

single project’s control portfolio. Finally,<br />

the detailed case analysis offers implications<br />

for global sourcing practices. The<br />

potentialtoreduceformalcontrolsoffers<br />

an important benefit for project managers<br />

of global, multisource, ISD out-<br />

• Transparency regarding roles and<br />

responsibilities of each party<br />

• Absence of negotiations and bargaining<br />

• Growing as a team<br />

• Development of joint project culture<br />

• Establishing an integrated service<br />

delivery process<br />

sourcing projects, as well as a likely reduction<br />

of the high management overhead<br />

costs normally associated with such<br />

projects.<br />

However, we also note several limitations.<br />

First, we did not have the opportunity<br />

to interview any representatives from<br />

one of the vendors, TEST, which was responsible<br />

for the software testing in a<br />

new factory in India. Second, the rampupphaseforthetestfactorydidnotwork<br />

out as planned, causing major problems<br />

for the project and its multiple vendors<br />

during the testing phase. In this case, the<br />

parties seemingly should have suffered<br />

great conflict, resulting from cultural differences.<br />

Surprisingly though, the cooperation<br />

was characterized by harmony,<br />

perhaps because each of the vendors had<br />

a prior business relationship with BANK.<br />

Further research should examine multisource,<br />

ISD outsourcing relationships<br />

in the context of newly composed vendor<br />

portfolios to examine this proposed<br />

explanation. Third, our results are specific<br />

to large, technology reengineering<br />

projects in the German financial services<br />

industry. Accordingly, we encourage researcherstocontinuetostudytheposed<br />

research question in other contexts and<br />

settings. Fourth, our analysis of the interplay<br />

of formal and informal management<br />

mechanisms and learning in global, multisource,<br />

ISD outsourcing projects could<br />

be extended to identify further aspects<br />

and reveal an increasingly differentiated<br />

picture of the interaction.<br />

Business & Information Systems Engineering


References<br />

Adler PS (2001) Market, hierarchy, and trust:<br />

the knowledge economy and the future<br />

of capitalism. Organization Science<br />

12(2):215–234<br />

Ang S, Inkpen AC (2008) Cultural intelligence<br />

and offshore outsourcing success: a framework<br />

of firm-level intercultural capability.<br />

Decision Sciences 39(3):337–358<br />

Bradach JL, Eccles RG (1989) Price, authority,<br />

and trust: from ideal types to plural forms.<br />

Annual Review of Sociology 15(1):97–118<br />

Cannon J, Achrol R, Gundlach G (2000) Contracts,<br />

norms, and plural form governance.<br />

Journal of the Academy of Marketing Science<br />

28(2):180–194<br />

Carmel E, Agarwal R (2001) Tactical approaches<br />

for alleviating distance in global<br />

software development. IEEE Software<br />

18(2):22–29<br />

Choudhury V, Sabherwal R (2003) Portfolios<br />

of control in outsourced software development<br />

projects. Information Systems<br />

Research 14(3):291–314<br />

Das TK, Teng B-S (1998) Between trust and<br />

control: developing confidence in partner<br />

cooperation in alliances. Academy of<br />

Management Review 23(3):491–512<br />

Das TK, Teng B-S (2001) Trust, control, and<br />

risk in strategic alliances: an integrated<br />

framework. Organization Studies 22(2):251<br />

David CG, Chand D, Newell S, Resende-Santos<br />

J (2008) Integrated collaboration across distributed<br />

sites: the perils of process and the<br />

promise of practice. Journal of Information<br />

Technology 23(1):44–54<br />

Dibbern J, Heinzl A, Leibbrandt S (2003) Interpretation<br />

des Sourcings der Informationsverarbeitung.<br />

Hintergründe und Grenzen<br />

ökonomischer Einflussgrößen. WIRT-<br />

SCHAFTSINFORMATIK 45(5):533–540<br />

Dibbern J, Winkler J, Heinzl A (2008) Explaining<br />

variations in client extra costs between<br />

software projects offshored to India. MIS<br />

Quarterly 32(2):333–366<br />

Gopal A, Sivaramakrishnan K (2008) On vendor<br />

preferences for contract types in offshoresoftwareprojects:thecaseoffixed<br />

price vs time and materials contracts. Information<br />

Systems Research 19(2):202–220<br />

Gregory R (2010a) The plural governance perspective:<br />

opportunities for is outsourcing<br />

research. In: Proc 18th European conference<br />

on information systems (ECIS 2010),<br />

Pretoria<br />

Gregory R (2010b) Review of the IS offshoring<br />

literature: the role of cross-cultural differences<br />

and management practices. In: Proc<br />

18th European conference on information<br />

systems (ECIS 2010), Pretoria<br />

Gupta U, Raval V (1999) Critical success factors<br />

for anchoring offshore projects. Information<br />

Strategy 15(2):21–27<br />

Hildenbrand T, Rothlauf F, Heinzl A (2007)<br />

Ansätze zur kollaborativen Softwareerstellung.<br />

WIRTSCHAFTSINFORMATIK<br />

49(Sonderheft):72–80<br />

Holmström Olsson H, Ó Conchúir E, Ågerfalk<br />

PJ, Fitzgerald B (2008) Two-stage offshoring:<br />

an investigation of the Irish bridge.<br />

MIS Quarterly 32(2):257–279<br />

Iacovou CL, Nakatsu R (2008) A risk profile<br />

of offshore-outsourced development<br />

projects. Communications of the ACM<br />

51(6):89–94<br />

Inkpen AC, Currall SC (2004) The coevolution<br />

of trust, control, and learning in joint ventures.<br />

Organization Science 15(5):586–599<br />

Kirsch LJ (1996) The management of complex<br />

tasks in organizations: controlling the sys-<br />

Business & Information Systems Engineering<br />

tems development process. Organization<br />

Science 7(1):1–21<br />

Kirsch LJ (2004) Deploying common systems<br />

globally: the dynamics of control. Information<br />

Systems Research 15(4):374–395<br />

Kliem R (2004) Managing the risks of offshore<br />

it development projects. Information<br />

Systems Management 21(3):22–27<br />

Kotlarsky J, Oshri I (2005) Social ties, knowledge<br />

sharing and successful collaboration<br />

in globally distributed system development<br />

projects. European Journal of Information<br />

Systems 14(1):37–48<br />

Kotlarsky J, van Fenema PC, Willcocks LP<br />

(2008) Developing a knowledge-based<br />

perspective on coordination: the case of<br />

global software projects. Information &<br />

Management 45(2):96–108<br />

Krishna S, Sahay S, Walsham G (2004) Managing<br />

cross-cultural issues in global software<br />

outsourcing. Communications of the ACM<br />

47(4):62–66<br />

Lacity MC, Willcocks LP (1998) An empirical<br />

investigation of information technology<br />

sourcing practices: lessons from experience.<br />

MIS Quarterly 22(3):363–408<br />

Lacity M, Khan SA, Willcocks L (2009) A review<br />

of the IT outsourcing literature: insights for<br />

practice. Journal of Strategic Information<br />

Systems 18(3):30–146<br />

Leonardi PM, Bailey DE (2008) Transformational<br />

technologies and the creation of<br />

new work practices: making implicit knowledge<br />

explicit in task-based offshoring. MIS<br />

Quarterly 32(2):411–436<br />

Levina N, Vaast E (2008) Innovating or doing<br />

as told? Status differences and overlapping<br />

boundaries in offshore collaboration. MIS<br />

Quarterly 32(2):307–332<br />

Meyer T, Stobbe A (2007) Welche “Standorte”<br />

wählen deutsche Unternehmen? WIRT-<br />

SCHAFTSINFORMATIK 49(Sonderheft):81–<br />

89<br />

Nicholson B, Sahay S (2001) Some political<br />

and cultural issues in the globalisation<br />

of software development: case experience<br />

from Britain and India. Information<br />

and Organization 11(1):25–43<br />

Nicholson B, Sahay S (2004) Embedded<br />

knowledge and offshore software development.<br />

Information and Organization<br />

14(4):329–365<br />

Oshri I, Kotlarsky J, Willcocks L (2007) Managing<br />

dispersed expertise in it offshore<br />

outsourcing: lessons from Tata consultancy<br />

services. MIS Quarterly Executive 6(2):53–<br />

65<br />

Rottman JW (2008) Successful knowledge<br />

transfer within offshore supplier networks:<br />

a case study exploring social capital in<br />

strategic alliances. Journal of Information<br />

Technology 23(1):31–43<br />

Stebbins RA (2001) Exploratory research in<br />

the social sciences. Sage, Thousand Oaks<br />

Tiwana A (2008) Does technological modularity<br />

substitute for control? A study<br />

of alliance performance in software outsourcing.<br />

Strategic Management Journal<br />

29(7):769–780<br />

Vlaar PWL, van Fenema PC, Tiwari V (2008)<br />

Cocreating understanding and value in distributed<br />

work: how members of onsite<br />

and offshore vendor teams give, make,<br />

demand, and break sense. MIS Quarterly<br />

32(2):227–255<br />

Vogt K, Gregory R, Beck R (2009) Measuring<br />

client-vendor distance in global<br />

outsourcing relationships: a conceptual<br />

model.In:Proc9.InternationaleTagung<br />

Wirtschaftsinformatik, Vienna<br />

Abstract<br />

Roman Beck, Katharina Schott<br />

BISE – RESEARCH PAPER<br />

The Interplay of Project Control<br />

and Interorganizational<br />

Learning: Mitigating Effects<br />

on Cultural Differences<br />

in Global, Multisource ISD<br />

Outsourcing Projects<br />

Research into global, multisource, information<br />

systems development outsourcing<br />

projects has uncovered management<br />

challenges, including cultural<br />

differences on multiple levels. While<br />

control mechanisms and interorganizational<br />

learning have been shown to<br />

contribute to the mitigation of cultural<br />

differences in such projects, a gap persists<br />

regarding the effect of the interplay<br />

between these mechanisms. This<br />

study employs an exploratory singlecase<br />

study design to analyze how formal<br />

and informal control mechanisms<br />

and interorganizational learning interact<br />

and thus contribute to the mitigation<br />

of cultural differences in global,<br />

multisource, information systems development<br />

outsourcing projects. With<br />

the key finding that the influence of informal<br />

controls and interorganizational<br />

learning on formal controls changes<br />

over time, this research helps expand<br />

the domain of control dynamics in global<br />

IS projects. This study also contributes<br />

to literature on ways to handle cultural<br />

differences in global, multisource, IS<br />

outsourcing projects.<br />

Keywords: Cultural differences, Formal<br />

control, Informal control, Interorganizational<br />

learning, IS outsourcing,<br />

Multisourcing, Global information systems<br />

development


BISE – RESEARCH PAPER<br />

Walsham G (2002) Cross-cultural software<br />

production and use: a structurational analysis.<br />

MIS Quarterly 26(4):359–380<br />

Willcocks LP, Lacity MC, Cullen S (2007) Information<br />

technology sourcing research:<br />

critique, lessons and prospects. In: Proc<br />

13th Americas conference on information<br />

systems, Keystone<br />

Winkler J, Dibbern J, Heinzl A (2007) Der<br />

Einfluss kultureller Unterschiede beim<br />

IT-Offshoring. WIRTSCHAFTSINFORMATIK<br />

49(2):95–103<br />

Yin R (2003) Case study research – design and<br />

methods. Sage, Thousand Oaks<br />

Business & Information Systems Engineering


Maximizing Cloud Provider <strong>Prof</strong>it from Equilibrium Price Auctions [PRE-PRINT]<br />

Ulrich Lampe, Melanie Siebenhaar, Apostolos Papageorgiou, Dieter Schuller, <strong>Ralf</strong> Steinmetz<br />

Multimedia Communications <strong>Lab</strong> (KOM)<br />

Technische Universität Darmstadt<br />

Darmstadt, Germany<br />

Email: ulrich.lampe@KOM.tu-darmstadt.de<br />

Abstract—Auctioning constitutes a market-driven scheme<br />

for the allocation of cloud-based computing capacities. It is<br />

practically applied today in the context of Infrastructure as a<br />

Service offers, specifically, virtual machines. However, the<br />

maximization of auction profits poses a challenging task for<br />

the cloud provider, because it involves the concurrent<br />

determination of equilibrium prices and distribution of<br />

virtual machine instances to the underlying physical hosts in<br />

the data center. In the work at hand, we propose an optimal<br />

approach, based on linear programming, as well as a<br />

heuristic approach to tackle this Equilibrium Price Auction<br />

Allocation Problem (EPAAP). Through an evaluation based on<br />

realistic data, we show the practical applicability and benefits<br />

of our contributions. Specifically, we find that the heuristic<br />

approach reduces the average computation time to solve an<br />

EPAAP by more than 99.9%, but still maintains a favorable<br />

average solution quality of 96.7% in terms of cloud provider<br />

profit, compared to the optimal approach.<br />

Keywords-Cloud Computing; IaaS; Equilibrium; Auction;<br />

Allocation; Optimization; Optimal; Heuristic<br />

A. Motivation<br />

I. INTRODUCTION<br />

In the past few years, cloud computing has gained<br />

tremendous interest both among practitioners and<br />

researchers. One of the essential ideas of this novel<br />

paradigm consists in the provision of Information<br />

Technology (IT) in a utility-like manner [1]. In the context<br />

of an envisioned cloud computing market, the decision<br />

whether to supply IT capacities in-house or lease them from<br />

the cloud is – aside from strategical considerations – largely<br />

determined by the price [2]. One potential instrument to<br />

achieve a market-based pricing and thus, efficient allocation<br />

of computing capacities, are auctions [3]. In this respect,<br />

Amazon Web Services 1 has not only been one of the<br />

pioneers in the cloud computing domain, but also among<br />

the first to employ auctions as a complement to traditional<br />

fixed-price schemes.<br />

Specifically, in the Spot Instances system 2 , cloud users<br />

submit bids to Amazon for Infrastructure as a Service (IaaS)<br />

offers in the form of Virtual Machines (VMs). Each bid<br />

states the desired number of instances and the maximum<br />

willingness to pay for a specific VM type. At a periodic<br />

1 http://aws.amazon.com/<br />

2 http://aws.amazon.com/ec2/spot-instances/<br />

time interval, Amazon determines an equilibrium price for<br />

each type. Users whose bids exceed (or meet) the<br />

equilibrium price are (partially) served with the desired<br />

instances. All users pay the identical equilibrium price –<br />

rather than their respective bid price – per instance.<br />

While many researchers have previously assumed that the<br />

price-setting mechanism is market-driven, i. e., determined<br />

by supply and demand [3]–[5], more recent research<br />

indicates that this is not the case. According to Agmon<br />

Ben-Yehuda et al. [6], the publicly posted equilibrium<br />

prices are likely selected at random from a small predefined<br />

interval. In the opinion of the aforementioned authors, this<br />

indicates a low demand in the Spot Instance system at<br />

present.<br />

However, it is safe to assume that with increasing<br />

acceptance and utilization of cloud computing offers,<br />

auctions will gain in popularity for the efficient allocation<br />

of computing capacities. Based on this notion, in the work<br />

at hand we examine how cloud providers can maximize<br />

their profit using equilibrium price auctions.<br />

B. Research Problem<br />

When applying equilibrium price auctions for the allocation<br />

of capacities, a cloud provider faces two distinct yet linked<br />

challenges. First, the provider has to determine specific<br />

equilibrium prices for each offered VM type. Second, the<br />

provider has to distribute the VM instances, which have<br />

been requested in the served (or satisfied) bids, among the<br />

Physical Machines (PM) instances in her/his data center. In<br />

this process, the cloud provider will commonly pursue the<br />

objective of maximizing profit. This profit is given by the<br />

difference between the revenue from the served bids and the<br />

operating costs of the PM instances. In the remainder of<br />

this paper, the combination of both challenges is referred to<br />

as Equilibrium Price Auction Allocation Problem (EPAAP).<br />

We assume that the cloud provider operates a limited<br />

number of PM types, which provide restricted resource<br />

supplies. These supplies are specified in terms of certain<br />

resource types, e. g., processor power or memory. Of each<br />

PM type, a restricted number of instances are available in<br />

the cloud provider’s data center(s). These PM instances may<br />

be independently powered on or off. Each active instance


leads to a fixed operating cost due to, e. g., idle power<br />

consumption and maintenance demands.<br />

We further assume that the cloud provider offers a<br />

predefined set of VM types, which exhibit certain resource<br />

demands. Due to these resource demands, each individual<br />

VM instance imposes additional variable operating costs on<br />

the PM instance that hosts it, due to, e. g., increased power<br />

consumption.<br />

Lastly, we assume that in accordance with the Spot<br />

Instances system, the cloud users may continuously submit<br />

or cancel bids. In contrast, the allocation process (i. e., the<br />

pricing of VM types and distribution of VM instances) is<br />

only periodically conducted by the cloud provider.<br />

The remainder of this paper is structured as follows: In<br />

Section II, we introduce a formal notation for the EPAAP<br />

and subsequently describe two optimization approaches, an<br />

optimal and a heuristic approach. Section III describes our<br />

evaluation procedure and the obtained results. In the<br />

subsequent Section IV, we provide an overview of related<br />

work. Lastly, Section V concludes the paper with a<br />

summary and outlook on future work.<br />

A. Formal Notations<br />

II. OPTIMIZATION APPROACHES<br />

As a basis for the optimization approaches that are presented<br />

in the following subsections, we introduce a formal notation<br />

for the EPAAP. First, we define the basic entities:<br />

• V ⊂ N: Set of offered VM types.<br />

• P ⊂ N: Set of available PM types.<br />

• R ⊂ N: Set of regarded resource types.<br />

Based on the previous definitions, the characteristics of and<br />

relations between the basic entities can be further specified:<br />

• RDvr ∈ R + : Resource demand of VM type v ∈ V for<br />

resource type r ∈ R.<br />

• RSpr ∈ R + : Resource supply of PM type p ∈ P for<br />

resource type r ∈ R.<br />

• CFp ∈ R + : Fixed operating cost of PM type p ∈ P<br />

per utilized instance.<br />

• CVpv ∈ R + : Variable operating cost of PM type<br />

p ∈ P per hosted instance of VM type v ∈ V .<br />

• np ∈ N: Available instances of PM type p ∈ P in the<br />

cloud provider’s data center(s).<br />

The bids that have been submitted by the cloud users are<br />

formalized using the following constructs:<br />

• B ⊂ N: Set of submitted bids.<br />

• Wb ∈ R + : Specified price for bid b ∈ B, i. e.,<br />

maximum willingness to pay.<br />

• Tb ∈ V : Specified VM type for bid b ∈ B.<br />

For reasons of simplicity and without loss of generality, we<br />

assume that each bid b ∈ B specifies an individual request<br />

for one VM instance. Thus, if a user bids for η ∈ N VM<br />

instances of the same type at an identical price – as it is the<br />

x111<br />

y11<br />

Bid B1 = B 1 1<br />

T1 = 1<br />

W1 = 0.20<br />

PM<br />

Instance<br />

P1, j = 1<br />

y12<br />

Bid B4 = B 1 2<br />

T4 = 1<br />

W4 = 0.15<br />

PM<br />

Instance<br />

P1, j = 2<br />

Bids for<br />

VM Type 1 (B 1 )<br />

PM Type 1 (P1)<br />

y21<br />

Bid B2 = B 2 1<br />

T2 = 2<br />

W2 = 0.25<br />

PM<br />

Instance<br />

P2, j = 1<br />

y22<br />

Bid B3 = B 2 2<br />

T3 = 2<br />

W3 = 0.20<br />

x4** x2** x3**<br />

x122<br />

x121<br />

x*11<br />

x112 x*12 x*21<br />

PM<br />

Instance<br />

P2, j = 2<br />

z11 z12 z21 z22<br />

x*22<br />

Bids for<br />

VM Type 2 (B 2 )<br />

Figure 1. Schematic overview of the optimization model, depicting the<br />

decision variables (in bold) and most relevant entities. A portion of the<br />

links has solely been sketched (in gray) to improve readability.<br />

case in the Spot Instances system – this results in η<br />

individual bids in the set B.<br />

Lastly, for reasons of convenience, we infer the following<br />

definitions from the previous specifications:<br />

• B v ⊆ B: Set of submitted bids for VM type v ∈ V .<br />

• W v ⊆ W : Set of prices for the bids in set B v .<br />

• mv ∈ N: Overall number of submitted bids for VM<br />

type v ∈ V .<br />

Without loss of generality, we establish that the bids in the<br />

set B are given in monotonically decreasing order of the<br />

corresponding prices. That is, it holds that Wb ≥ Wb ′ for<br />

b, b ′ ∈ B, b < b ′ . The same applies for the respective<br />

subsets of bids, i. e., Bv .<br />

B. Optimal Allocation Approach<br />

To compute an optimal solution to the EPAAP, we transfer<br />

the problem definition from Subsection I-B into a<br />

mathematical optimization model. The result is given in<br />

Model 1 and will be explained in detail in the following. In<br />

order to promote an easier interpretation of the model, we<br />

first discuss the significance of the decision variables.<br />

In Equation 12, x, y, and z are defined as binary decision<br />

variables. x is the primary decision variable, whereas y and<br />

z can be considered auxiliary decision variables.<br />

Specifically, xbpj indicates whether the bid b has been<br />

served and assigned to a PM instance of type p with the<br />

running index j or not. yvk indicates whether for a VM<br />

type v, exactly the first k bids are served or not. Lastly, zpj<br />

represents whether a PM instance of type p with the<br />

running index j is utilized, i. e., powered on, or not. An<br />

overview of the optimization model, which highlights the<br />

relations between the decision variables and the most<br />

important entities, is depicted in Figure 1.<br />

Equation 1 specifies the objective of the optimization model,<br />

namely the maximization of profit. The profit comprises<br />

PM Type 2 (P2)


three components. The first component is the revenue that is<br />

generated through the served bids. For that matter, yvk also<br />

indicates whether the k-th bid for VM type v corresponds<br />

to the equilibrium price or not. The second component is<br />

the fixed operating cost of all utilized PM instances, while<br />

the third component represents the additional variable<br />

operating cost due to the hosted VM instances.<br />

Equations 2 and 3 link the decision variables x and y and x<br />

and z respectively. Equation 4 assures that the resource<br />

demands of all VM instances are met by the resource<br />

supplies of the respective PM instance that host them.<br />

Equation 5 guarantees that an individual bid cannot be<br />

served more than once. Equation 6 ensures that solely one<br />

equilibrium price for each VM type can be set.<br />

Equation 7, for performance reasons, restricts the solution<br />

space by preferring PM instances of the same type with a<br />

lower running index over those with a higher running index.<br />

Equation 8, also for performance reasons, excludes<br />

dominated solutions from the solution space. A solution is<br />

referred to as dominated if another solution exists that<br />

would yield a higher or equal revenue, even if a smaller<br />

number of bids for a specific VM type is served.<br />

Equations 9 through 11 define valid running indices for VM<br />

and PM instances. The definition of the latter is based on<br />

two observations: First, the number of utilized PM instances<br />

cannot exceed the available number of instances of this<br />

type, np. Second, in the theoretical case that all bids were<br />

served and the requested VM instances were each assigned<br />

to an individual PM instance of the same type, no more<br />

than |B| instances would be required.<br />

As can be seen, Model 1 constitutes a Linear Program (LP),<br />

or more specifically, Binary Integer Program (BIP). This<br />

class of optimization problems can be solved using<br />

well-known methods from the field of Operations Research,<br />

most notably, the Branch and Bound (B&B) algorithm [7].<br />

While the B&B algorithm can be very efficient in some<br />

cases, it is still based on the principle of enumeration, i. e.,<br />

in the worst case, all potential solutions have to be<br />

examined [8].<br />

Specifically, for a BIP, the solution space grows<br />

exponentially with the number of decision variables. As can<br />

be observed from Model 1 (notably, Equations 1 and 12),<br />

the number of decision variables increases quadratically<br />

with the number of bids and linearly with the number of<br />

PM types. Accordingly, the computational complexity of the<br />

optimal allocation approach is exponential and corresponds<br />

to O(2 |B|2 ∗|P | ).<br />

C. Heuristic Allocation Approach<br />

For real-life application scenarios involving thousands of<br />

bids, the optimal allocation approach may be problematic<br />

due to its exponential growth in computational complexity.<br />

Thus, we have developed a heuristic approach that trades<br />

Model 1 Optimal Allocation Model<br />

Maximize P r(x, y, z) = �<br />

− �<br />

p∈P,j∈Jp<br />

k ∗ yvk ≤<br />

zpj ∗ CFpj −<br />

�<br />

1≤i≤k,p∈P,j∈Jp<br />

v∈V,k∈Kv<br />

�<br />

b∈B,p∈P,j∈Jp<br />

subject to<br />

yvk ∗ k ∗ W v k<br />

xbpj ∗ CVpTb<br />

(1)<br />

xB v i pj ∀v ∈ V, k ∈ Kv (2)<br />

zpj ≥ xbpj ∀b ∈ B, p ∈ P, j ∈ Jp (3)<br />

�<br />

xbpj ∗ RDTbr ≤ RSpr ∀p ∈ P, j ∈ Jp, r ∈ R (4)<br />

b∈B<br />

Jp =<br />

�<br />

p∈P,j∈Jp<br />

�<br />

k∈Kv<br />

xbpj ≤ 1 ∀b ∈ B (5)<br />

yvk ≤ 1 ∀v ∈ V (6)<br />

zpj ≥ zpj ′ ∀p ∈ P, j ∈ Jp, j ′ ∈ Jp, j < j ′<br />

yvk = 0 if k ∗ W v k ≤ k ′ ∗ W v k ′<br />

∀k ∈ Kv, k ′ ∈ Kv, k > k ′<br />

Kv =<br />

�<br />

�<br />

{1, ..., mv}<br />

∅<br />

if mv > 0<br />

else<br />

{1, ..., min(np, |B|)} if np > 0<br />

∅ else<br />

(7)<br />

(8)<br />

∀v ∈ V (9)<br />

∀p ∈ P (10)<br />

Jp, Kv ⊂ N (11)<br />

xbpj ∈ {0, 1} ∀b ∈ B, p ∈ P, j ∈ Jp<br />

yvk ∈ {0, 1} ∀v ∈ V, k ∈ Kv (12)<br />

zpj ∈ {0, 1} ∀p ∈ P, j ∈ Jp<br />

reductions in computation time against potentially<br />

sub-optimal solutions.<br />

The principle idea is to initially determine the equilibrium<br />

prices for all VM types, such that the expected profit from<br />

the served bids is maximized. Subsequently, these served<br />

bids – or more specifically, the VM instances that have<br />

been requested in these bids – are cost-efficiently distributed<br />

across the physical hosts. Accordingly, the approach is split<br />

into two phases, VM pricing and VM distribution.<br />

The procedure for the second phase is inspired by a<br />

heuristic for the distribution of software services across VM<br />

instances, which we have introduced in our previous work


[9]. This heuristic, in turn, adapts concepts that are<br />

frequently applied for solving the well-known Knapsack<br />

problem [7]. For additional details, we refer to our previous<br />

publication.<br />

The pseudo code for the two phases is provided in Listing 1<br />

and 2. In accordance with the previous subsection, xbpj<br />

represents the main binary decision variable. Qv ∈ R +<br />

denotes the equilibrium price for each VM type v. Lastly,<br />

the set S ⊆ B contains the bids that will be served.<br />

In the first phase, VM pricing, we initially estimate the cost<br />

of serving an individual bid for each VM type v. This<br />

serving cost CSvp of a VM type v on each PM type p<br />

corresponds to the partial fixed and additional variable<br />

operating cost of a respective PM instance. The partial fixed<br />

operating cost, in this context, is the fixed operating cost of<br />

a PM instance, multiplied by the ratio between the resource<br />

demands of VM type v and the resource supplies of PM<br />

type p. The individual serving costs are aggregated into an<br />

weighted average serving cost CSv across all suitable PM<br />

types for each VM type v, using the respective weights gvp.<br />

Suitable, in this respect, means that a PM type can host a<br />

VM type subject to the given resource constraints (lines<br />

3-13). On the basis of these serving costs and the initial bid<br />

prices, we determine a favorable number ˆ kv ∈ N of bids for<br />

each VM type v that should be served. The number is<br />

considered favorable if the expected profit Ev ∈ R, i. e., the<br />

difference between the revenue from the bids and the<br />

expected serving costs, becomes maximal. In the same step,<br />

the equilibrium price Qv for each VM type v is inferred.<br />

The served bids are stored in the set S, which, in<br />

accordance with the set B, we assume to be ordered by<br />

monotonically decreasing bid prices (lines 14-21). The<br />

corresponding VM instances are considered for distribution<br />

in the following phase.<br />

In the second phase, VM distribution, we first initialize an<br />

instance count jp ∈ N for each PM type p ∈ P (lines 1-3).<br />

Subsequently, the following packing process is conducted:<br />

We scan the list of PM types to determine a favorite ˆp ∈ P .<br />

For this purpose, we initially create a packing list Lp ⊆ S<br />

for each type p, unless the maximum number of instances<br />

np of this type has already been reached. A packing list<br />

constitutes a set of bids that could be hosted by a new<br />

instance of the regarded PM type. For that matter, we scan<br />

the list of served bids S. For each bid b, we check whether<br />

it could – in addition to the current packing list Lp – be<br />

hosted by a new instance of type p, based on the given<br />

resource constraints. If the bid b meets this condition, it is<br />

added to the packing list. If not, the associated VM type Tb<br />

is added to an ignore list Ip ⊆ V , and bids of the identical<br />

type are ignored during the remainder of the packing list<br />

creation (lines 8-17). In the following, we compute the<br />

utility Up of the current PM type p. Utility is defined as the<br />

ratio between the revenue from the packing list Lp and the<br />

resulting fixed and variable operating costs of a new PM<br />

Listing 1 Algorithm for VM Pricing<br />

1: S ← ∅<br />

2: for all v ∈ V do<br />

3: gv ← 0, CSv ← 0<br />

4: for all p ∈ P do<br />

5: if p.canHost(v) = true then<br />

6: gvp ← min(np, |B|)<br />

7:<br />

8:<br />

gv ← gv + gvp<br />

CSvp ← CVpv + CFp ∗ 1<br />

9:<br />

�<br />

|R| r∈R<br />

CSv ← CSv + gvp ∗ CSvp<br />

10: end if<br />

11: end for<br />

12: if gv > 0 then<br />

13:<br />

14:<br />

CSv ← CSv/gv<br />

kv<br />

ˆ ← 0, Êv ← 0, Qv ← ∞<br />

15:<br />

16:<br />

for k = 1 → mv do<br />

Ev ← k ∗ (W v 17:<br />

k − CSv)<br />

if Ev > Êv then<br />

18: Êv ← Ev, ˆ kv ← k, Qv ← W v 19:<br />

k<br />

end if<br />

20: end for<br />

21: S ← S ∪ {Bv 1 , ..., Bv 22:<br />

kv ˆ }<br />

end if<br />

23: end for<br />

� RDvr<br />

RSpr<br />

instance. If a PM type exhibits higher utility than the<br />

current favorite ˆp, it is stored as new favorite (lines 18-21).<br />

After all PM types have been scanned and a favorite has<br />

been identified, all bids in the packing list Lˆp are assigned<br />

to a new instance of type ˆp with the previously incremented<br />

index jˆp. The bids are also removed from the set S (lines<br />

26-29). The packing process is repeated until all served bids<br />

have been assigned or no suitable (i. e., favorite) PM<br />

instance remains for assignment.<br />

In terms of computational complexity, the heuristic has<br />

substantial advantages over the optimal approach: As it can<br />

be observed from Listing 1, the computational complexity<br />

of the first phase grows linearly with the number of PM<br />

types and VM types. The latter is commonly substantially<br />

smaller than the number of bids and thus negligible. For the<br />

second phase, according to Listing 2, the maximum number<br />

of iterations corresponds to the number of bids. In each<br />

iteration, the computational complexity relates linearly to<br />

the number of bids and PM types again. Thus, the<br />

computational complexity of the heuristic allocation<br />

approach is polynomial and corresponds to O(|B| 2 ∗ |P |).<br />

III. EVALUATION<br />

Both optimization approaches from the previous section<br />

have been implemented in a prototypical Java program. In<br />

order to solve the optimization model for the first allocation<br />

approach, we map it into a programmatic representation<br />


Listing 2 Algorithm for VM Distribution<br />

1: for all p ∈ P do<br />

2: jp ← 0<br />

3: end for<br />

4: repeat<br />

5: ˆp ← ∅, Û ← 0<br />

6: for all p ∈ P do<br />

7: if jp < np then<br />

8: Lp ← ∅, Ip ← ∅<br />

9: for all b ∈ S do<br />

10: if Tb /∈ Ip then<br />

11: if p.canHost(Lp ∪ b) = true then<br />

12: Lp ← Lp ∪ b<br />

13: else<br />

14: Ip ← Ip ∪ Tb<br />

15: end if<br />

16: end if<br />

17: end for<br />

18: Up ←<br />

� �<br />

b∈Lp QTb<br />

19:<br />

20:<br />

if Up > Û then<br />

Û ← Up, ˆp ← p<br />

21: end if<br />

22: end if<br />

23: end for<br />

24: if ˆp �= ∅ then<br />

25: jˆp ← jˆp + 1<br />

26: for all b ∈ Lˆp do<br />

27:<br />

28:<br />

xbvj ← 1 ˆp<br />

end for<br />

29: S ← S \ Lˆp<br />

30: end if<br />

31: until S = ∅ ∨ ˆp = ∅<br />

� �<br />

/ CFp + �<br />

b∈Lp CVpTb<br />

�<br />

using the JavaILP framework 3 . As actual solvers, the<br />

commercial IBM CPLEX Optimizer 4 and the free lpsolve 5<br />

frameworks may be employed, with the first constituting the<br />

default choice in our evaluation.<br />

A. Approach and Methodology<br />

The aim of our evaluation lies in a quantitative assessment<br />

of the two optimization approaches. Thus, the evaluation<br />

complements the brief and solely qualitative analyses of<br />

computational complexity from Subsections II-B and II-C.<br />

Our focus lies on two metrics that are of practical relevance<br />

in the context of auction-based VM allocation: First, the<br />

metric computation time demonstrates the overall scalability<br />

of the approaches. It also expresses the delay that is<br />

introduced into the allocation process through the<br />

3 http://javailp.sourceforge.net/<br />

4 http://www.ibm.com/software/integration/optimization/cplex-optimizer/<br />

5 http://sourceforge.net/projects/lpsolve/<br />

application of the optimization approaches. Second, the<br />

metric profit represents the absolute solution quality of the<br />

approaches. It thus expresses the utility of an optimized<br />

allocation to the cloud provider in monetary terms.<br />

For the evaluation, we have created 18 distinct classes of<br />

EPAAPs. Each class contains 100 individual problems.<br />

Across the different classes, we vary the problem dimension<br />

with respect to the regarded number of bids and PM types.<br />

The number of VM types and resource types are fixed<br />

across all classes. Each problem is randomly generated<br />

based on realistic data, which has been obtained as<br />

described in the following.<br />

For the data of the VM types in the evaluation, we use the<br />

specifications provided by Amazon Web Services for its<br />

Elastic Computing Cloud (EC2) and Spot Instances offers 6 .<br />

Excluding special-purpose and non-deterministic types<br />

(namely, Cluster and Micro), we infer eight different VM<br />

types. Each of these types exhibits specific resource<br />

demands with respect to three resource types, namely<br />

processor (CPU), memory (RAM), and storage (HDD).<br />

Unfortunately, to the best of our knowledge, Amazon Web<br />

Services has not published detailed information about the<br />

PMs in its data centers to date. However, empirical results<br />

by an industry researcher indicate that the most resource<br />

intensive VM types are run on dedicated PMs [10]. Based<br />

on this notion, the specifications of the High-Memory<br />

Quadruple Extra Large VM type, which exhibits the highest<br />

demands for each resource type, is assumed as baseline for<br />

the definition of five different PM types. Based on<br />

calculations by Walker [2], the fixed operating costs of<br />

these PM types are conservatively estimated to range from<br />

$0.20 to $0.40 per hour. The specific figure for each PM<br />

type is correlated with its resource supply.<br />

The bid prices that the cloud users submit to the cloud<br />

provider were modeled using specific distribution functions<br />

Fv for each VM type v. For the choice of these distribution<br />

functions, we reason as follows: The lowest observed bid<br />

price, αv, will most likely correspond to the lowest<br />

permissible bid price, i. e., αv = 0.01. Given that the users<br />

act rational, the highest observed bid price, γv, will not<br />

exceed the price Ov ∈ R of a so-called On-Demand VM<br />

instance, which imposes a static usage fee, but essentially<br />

offers guaranteed availability in return. Thus, we assume<br />

γv = Ov − 0.01. Lastly, according to empirical findings by<br />

Wee [5] and statements by Amazon Web Services [11], the<br />

average savings from Amazon Spot Instances – compared to<br />

On-Demand instances – amount to 52.3% and between 50%<br />

and 66% respectively. Thus, we assume that the most<br />

frequently observed bid price, βv, corresponds to 50% of<br />

the On-Demand price, i. e., βv = 0.5 ∗ Ov. The underlying<br />

notion is that the cloud users would like to maximize their<br />

savings from the auction mechanism, but also maintain a<br />

6 http://aws.amazon.com/en/ec2/#instance


easonable chance of actually being allocated VM instances.<br />

Based on the previous reasoning, we believe that the use of<br />

a triangle distribution, i. e., Fv ∼ T r(αv, βv, γv), best<br />

reflects a realistic bidding behavior.<br />

Based on past research by Greenberg et al. [12] and Barroso<br />

and Hölzle [13], we estimate that the additional variable<br />

operating cost of a fully utilized server amounts to about<br />

25% of its fixed operating cost. As a first approximation, we<br />

further assume that the cost of hosting a certain VM type<br />

linearly relates to the resource utilization on the underlying<br />

PM. Accordingly, the variable operating costs of each VM<br />

type on every PM type are determined using Equation 13.<br />

CVpv = 0.25∗CFp∗ 1<br />

|R| ∗�<br />

RDvr<br />

RSpr<br />

r∈R<br />

B. Results and Discussion<br />

∀p ∈ P, v ∈ V (13)<br />

Following the problem generation, each EPAAP was solved<br />

using both optimization approaches, optimal and heuristic.<br />

In the process, the respective computation time and<br />

resulting profit for the cloud provider were recorded. In<br />

order to complete the evaluation within an acceptable<br />

amount of time, a timeout period of 5 minutes (i. e., 300<br />

seconds) was imposed per problem and approach. The<br />

evaluation was conducted on a desktop computer with an<br />

Intel Core 2 Quad Q9450 processor and 4 GB of memory,<br />

operating under Microsoft Windows 7.<br />

Table I provides an overview of the evaluated EPAAP<br />

classes. As it has been previously outlined, the number of<br />

bids (dB) and PM types (dP ) was varied for each problem<br />

class, whereas the number of VM types and resource types<br />

was fixed (dV = 8, dR = 3). The table also indicates the<br />

percentage of problems that could be solved within the<br />

specified timeout period by each optimization approach.<br />

Only those problems that could be solved by both<br />

approaches served as sample for the evaluation results<br />

provided hereafter. Please note that problem classes E and<br />

F constitute an exception, because they were exclusively<br />

solved using the heuristic approach in order to show its<br />

applicability to large-scale problems.<br />

To begin with, Figure 2 depicts the absolute average<br />

computation times per problem across all considered<br />

problem classes. In accordance with the qualitative<br />

discussion, the computation times for the optimal approach<br />

grow roughly exponentially with the number of bids. The<br />

effect of an increasing number of PM types is less<br />

accentuated, but still well observable. For the problem<br />

classes involving 20 bids (B1−4), the absolute computation<br />

time of the optimal approach reaches the magnitude order<br />

of one second. For problem classes that involve 30 bids or<br />

more (C1−4 and D1−4), the computation time reaches and<br />

even exceeds the magnitude order of ten seconds.<br />

Accordingly, a substantial share of problems cannot be<br />

solved at all within the specified timeout period (cf.<br />

Table I<br />

OVERVIEW OF EVALUATED EPAAP CLASSES AND SHARE OF SOLVED<br />

PROBLEMS PER CLASS.<br />

Problem class Solved problems (%)<br />

ID dP dB Opt. Heur. Both<br />

A1 1 10 100 100 100<br />

A2 1 20 100 100 100<br />

A3 1 30 100 100 100<br />

A4 1 40 100 100 100<br />

B1 2 10 100 100 100<br />

B2 2 20 100 100 100<br />

B3 2 30 100 100 100<br />

B4 2 40 100 100 100<br />

C1 3 10 95 100 95<br />

C2 3 20 97 100 97<br />

C3 3 30 85 100 85<br />

C4 3 40 80 100 80<br />

D1 4 10 84 100 84<br />

D2 4 20 72 100 72<br />

D3 4 30 54 100 54<br />

D4 4 40 46 100 46<br />

E 4 1000 – 100 –<br />

F 4 10000 – 100 –<br />

Table I). In contrast, for the heuristic approach, the average<br />

computation time does not exceed the magnitude order of<br />

one millisecond up to problem class D. For problem classes<br />

E and F , the average computation time lies in the<br />

magnitude order of ten milliseconds and one second<br />

respectively. In general, the increase with growing problem<br />

size is rather moderate and roughly corresponds to the<br />

square of the number of bids. In accordance, the timeout is<br />

not of practical relevance to the heuristic approach.<br />

The performance difference between the two approaches is<br />

further highlighted by Figure 3, which depicts the ratio of<br />

computation time between the heuristic and optimal<br />

approach. For the smallest problem classes involving 10<br />

bids (A1−4), the ratio amounts to approximately 0.2% or<br />

less. This is equivalent to a reduction of more than 99.8% in<br />

computation time. For the larger problem classes involving<br />

additional bids, the ratio converges toward 0, indicating<br />

reductions of more than 99.9% by the heuristic approach.<br />

The results for the second metric, profit, are given in<br />

Figure 4. As can be seen, the heuristic approach achieves<br />

favorable and consistent results, ranging between about<br />

95.2% and 97.6% compared to the optimal solution. On<br />

average across all classes, the figure corresponds to<br />

approximately 96.7%. That means, due to the application of<br />

the heuristic allocation approach, the cloud provider would<br />

incur an average reduction in profit of 3.3% compared to<br />

the optimal solution.<br />

In summary, the results indicate that the computation of an<br />

optimal solution to the EPAAP is difficult to achieve under<br />

practical conditions. A cloud provider will usually receive


Computation time [ms]<br />

100000<br />

10000<br />

1000<br />

100<br />

10<br />

1<br />

0.1<br />

A1:1;10<br />

A2:2;10<br />

A3:3;10<br />

A4:4;10<br />

B1:1;20<br />

B2:2;20<br />

B3:3;20<br />

B4:4;20<br />

C1:1;30<br />

C2:2;30<br />

C3:3;30<br />

C4:4;30<br />

EPAAP class (ID:dP ;dB)<br />

Optimal approach<br />

Heuristic approach<br />

Figure 2. Mean absolute computation times per problem for both<br />

optimization approaches.<br />

Ratio of computation time [%]<br />

0.25<br />

0.2<br />

0.15<br />

0.1<br />

0.05<br />

0<br />

A1:1;10<br />

A2:2;10<br />

A3:3;10<br />

A4:4;10<br />

B1:1;20<br />

B2:2;20<br />

B3:3;20<br />

B4:4;20<br />

C1:1;30<br />

C2:2;30<br />

C3:3;30<br />

C4:4;30<br />

EPAAP class (ID:dP ;dB)<br />

D1:1;40<br />

D1:1;40<br />

D2:2;40<br />

D3:3;40<br />

D2:2;40<br />

D4:4;40<br />

Heuristic / Optimal approach<br />

Figure 3. Ratios of computation times between both optimization<br />

approaches (based on micro-average).<br />

Ratio of profit [%]<br />

100<br />

98<br />

96<br />

94<br />

92<br />

90<br />

A1:1;10<br />

A2:2;10<br />

A3:3;10<br />

A4:4;10<br />

B1:1;20<br />

B2:2;20<br />

B3:3;20<br />

B4:4;20<br />

C1:1;30<br />

C2:2;30<br />

C3:3;30<br />

C4:4;30<br />

EPAAP class (ID:dP ;dB)<br />

D1:1;40<br />

D2:2;40<br />

D3:3;40<br />

D3:3;40<br />

E:4;1000<br />

D4:4;40<br />

Heuristic / Optimal approach<br />

Figure 4. Ratios of profits between both optimization approaches (based<br />

on micro-average).<br />

D4:4;40<br />

F:4;10000<br />

- All -<br />

- All -<br />

thousands of bids that have to be regarded in the allocation<br />

process. At the same time, the cloud provider underlies<br />

stringent time constraints, given that the timespan between<br />

accepting the last bids in an auction period and the<br />

announcement of the resulting equilibrium prices and VM<br />

allocations should be minimal. For such application<br />

scenarios, our proposed heuristic approach presents a viable<br />

option. It achieves substantial reductions in computation<br />

time, but also retains a favorable solution quality in terms<br />

of profit for the cloud provider.<br />

IV. RELATED WORK<br />

To the best of our knowledge, we are the first to<br />

scientifically examine the Equilibrium Price Auction<br />

Allocation Problem, i. e., the challenge of concurrently<br />

pricing and distributing VM instances based on an auction<br />

scheme. However, in the broader context of cloud<br />

computing, a substantial amount of work has been<br />

conducted with respect to related topics. In the following,<br />

we focus on a set of papers that we consider representative<br />

for each topic area.<br />

Breitgand et al. [14], for instance, have proposed an<br />

optimization model and corresponding heuristic for the<br />

distribution of VM instances on physical hosts through a<br />

cloud provider. The authors take into account various<br />

constraints, including resource demands and supplies, and<br />

also permit the definition of different objectives, including<br />

profit maximization. However, their research does not<br />

address the aspect of pricing the VMs in an auction-based<br />

setting.<br />

Korupolu et al. [15] have proposed an optimization scheme<br />

for the placement of applications, which comprise compute<br />

and storage components, in data centers. In accordance with<br />

our work, their heuristic approach is inspired by the<br />

Knapsack problem. However, the work of Korupolu et al.<br />

does not involve an auction-based pricing mechanism.<br />

Zaman and Grosu [3] have examined the allocation of VM<br />

instances to physical machines based on combinatorial<br />

auctions, where cloud users bid for arbitrary bundles of VM<br />

instances. The authors propose multiple heuristic allocation<br />

strategies, which are evaluated with respect to different<br />

objectives, including maximization of revenues. In contrast<br />

to our work, the prices of identical VM types may be<br />

discriminated between different users, which is not the case<br />

in equilibrium price auctions. Furthermore, the authors do<br />

not explicitly regard the distribution among physical hosts<br />

under resource constraints.<br />

Zaman and Grosu [16] have additionally addressed the issue<br />

of auction-based VM allocation in a more recent paper.<br />

However, the focus of this work lies on bidding strategies<br />

for the cloud user, rather than optimization approaches for<br />

the cloud provider.<br />

Lin et al. [17], in accordance with our work, have proposed<br />

an allocation mechanism for second-price (i. e., equilibrium


price) auctions. However, the authors solely focus on the<br />

optimal pricing of resources, but do not consider the<br />

concurrent distribution of VM instances across physical<br />

hosts.<br />

Özer and Özturan [18] have proposed an optimal approach,<br />

as well as different heuristics for the allocation of grid and<br />

cloud resources. In contrast to our work, the authors assume<br />

combinatorial auctions, where users submit bids for bundles<br />

of resources, which results in a different pricing approach.<br />

In addition, Özer and Özturan do not consider the<br />

distribution to physical hosts as part of the allocation<br />

process.<br />

In our own previous research [9], we have examined the<br />

Software Service Distribution Problem. This challenge<br />

concerns the cost-minimal distribution of Software as a<br />

Service instances across leased VM instances under<br />

resource constraints. We have presented an optimal, as well<br />

as heuristic solution approach. However, these approaches<br />

only address the distribution process. They do not cover the<br />

concurrent pricing of entities, which constitutes a major<br />

challenge in the EPAAP.<br />

V. SUMMARY AND OUTLOOK<br />

In the work at hand, we have introduced the Equilibrium<br />

Price Auction Allocation Problem (EPAAP), a challenge in<br />

the context of cloud computing. This problem concerns<br />

cloud providers and involves the concurrent pricing and<br />

distribution of virtual machines across physical machines<br />

based on equilibrium price auctions.<br />

As first major contribution, we have introduced a<br />

mathematical formulation of the EPAAP as binary integer<br />

program. This model serves as the basis of an optimal<br />

allocation approach, which permits the computation of<br />

profit-maximizing solutions to the EPAAP. As second major<br />

contribution, given the computational complexity of the<br />

optimal approach, we have developed a heuristic approach.<br />

This heuristic trades substantial reductions in computation<br />

time against small reductions in overall profit.<br />

Through an evaluation based on realistic data from the<br />

cloud computing domain, we have demonstrated the<br />

practical applicability of our approaches. Specifically, we<br />

have shown that the heuristic is able to achieve reductions<br />

in computation time of more than 99.9% compared to an<br />

optimal approach. At the same time, it achieves profits that<br />

correspond to about 96.7% of the optimal solution on<br />

average. Thus, our work is not only of scientific interest, but<br />

can also provide a foundation for the practical application of<br />

equilibrium price auctions in the cloud computing domain.<br />

In our future work, we will aim at the further improvement<br />

of the heuristic approach with respect to the solution quality.<br />

In addition, we plan to extend the proposed approaches<br />

such that the specific requirements of an optimization across<br />

multiple subsequent auction periods, such as the live<br />

migration of virtual machines, are supported.<br />

ACKNOWLEDGMENTS<br />

This work has partly been sponsored by E-<strong>Finance</strong> <strong>Lab</strong><br />

e. V., Frankfurt a.M., Germany (www.efinancelab.de).<br />

REFERENCES<br />

[1] R. Buyya, C. Yeo, S. Venugopal, J. Broberg, and I. Brandic,<br />

“Cloud Computing and Emerging IT Platforms: Vision, Hype,<br />

and Reality for Delivering Computing as the 5th Utility,”<br />

Future Generation Computer Systems, vol. 25, no. 6, pp.<br />

599–616, 2009.<br />

[2] E. Walker, “The Real Cost of a CPU Hour,” Computer,<br />

vol. 42, no. 4, pp. 35–41, 2009.<br />

[3] S. Zaman and D. Grosu, “Combinatorial Auction-Based<br />

Allocation of Virtual Machine Instances in Clouds,” in IEEE<br />

2nd Int. Conf. on Cloud Computing Technology and Science,<br />

2010.<br />

[4] R. T. Bahman Javadi and R. Buyya, “Statistical Modeling of<br />

Spot Instance Prices in Public Cloud Environments,” in 4th<br />

IEEE/ACM Int. Conf. on Utility and Cloud Computing, 2011.<br />

[5] S. Wee, “Debunking Real-Time Pricing in Cloud Computing,”<br />

in 11th IEEE/ACM Int. Symp. on Cluster, Cloud and Grid<br />

Computing, 2011.<br />

[6] O. Agmon Ben-Yehuda, M. Ben-Yehuda, A. Schuster, and<br />

D. Tsafrir, “Deconstructing Amazon EC2 Spot Instance<br />

Pricing,” in IEEE 3rd Int. Conf. on Cloud Computing<br />

Technology and Science, 2011.<br />

[7] W. Domschke and A. <strong>Dr</strong>exl, Einführung in Operations<br />

Research, 6th ed. Springer, 2004, in German.<br />

[8] F. Hillier and G. Lieberman, Introduction to Operations<br />

Research, 8th ed. McGraw-Hill, 2005.<br />

[9] U. Lampe, T. Mayer, J. Hiemer, D. Schuller, and<br />

R. Steinmetz, “Enabling Cost-Efficient Software Service<br />

Distribution in Infrastructure Clouds at Run Time,” in 2011<br />

IEEE Int. Conf. on Service Oriented Computing &<br />

Applications, 2011.<br />

[10] H. Liu, “Amazon’s Physical Hardware and EC2 Compute<br />

Unit,” http://huanliu.wordpress.com/2010/06/14/<br />

amazons-physical-hardware-and-ec2-compute-unit/, last<br />

accessed May 10, 2012.<br />

[11] Amazon Web Services LLC, “Using Amazon EC2 Spot<br />

Instances for Scientific Computing,”<br />

http://aws.amazon.com/en/ec2/spot-and-science/, last accessed<br />

May 10, 2012.<br />

[12] A. Greenberg, J. Hamilton, D. Maltz, and P. Patel, “The Cost<br />

of a Cloud: Research Problems in Data Center Networks,”<br />

ACM SIGCOMM Computer Communication Review, vol. 39,<br />

no. 1, pp. 68–73, 2008.<br />

[13] L. Barroso and U. Hölzle, “The Case for Energy-Proportional<br />

Computing,” Computer, vol. 40, no. 12, pp. 33–37, 2007.<br />

[14] D. Breitgand, A. Maraschini, and J. Tordsson, “Policy-<strong>Dr</strong>iven<br />

Service Placement Optimization in Federated Clouds (TR<br />

H-0299),” IBM Research, Tech. Rep., 2011.


[15] M. Korupolu, A. Singh, and B. Bamba, “Coupled Placement<br />

in Modern Data Centers,” in 2009 IEEE Int. Symp. on<br />

Parallel & Distributed Processing, 2009.<br />

[16] S. Zaman and D. Grosu, “Efficient Bidding for Virtual<br />

Machine Instances in Clouds,” in 2011 IEEE Int. Conf. on<br />

Cloud Computing, 2011.<br />

[17] W. Lin, G. Lin, and H. Wei, “Dynamic Auction Mechanism<br />

for Cloud Resource Allocation,” in 10th IEEE/ACM Int. Conf.<br />

on Cluster, Cloud and Grid Computing, 2010.<br />

[18] A. Özer and C. Özturan, “An Auction Based Mathematical<br />

Model and Heuristics for Resource Co-Allocation Problem in<br />

Grids and Clouds,” in 5th Int. Conf. on Soft Computing,<br />

Computing with Words and Perceptions in System Analysis,<br />

Decision and Control, 2009.


Cost-driven Optimization of Complex Service-based Workflows<br />

for Stochastic QoS Parameters<br />

Dieter Schuller, Ulrich Lampe, Julian Eckert, <strong>Ralf</strong> Steinmetz<br />

Multimedia Communications <strong>Lab</strong> (KOM)<br />

Technische Universität Darmstadt<br />

Darmstadt, Germany<br />

Email: {firstname.lastname}@KOM.tu-darmstadt.de<br />

Abstract—The challenge of optimally selecting services from<br />

a set of functionally appropriate ones under Quality of Service<br />

(QoS) constraints – the Service Selection Problem – has been<br />

extensively addressed in the literature based on deterministic<br />

parameters. In practice, however, Quality of Service QoS<br />

parameters rather follow a stochastic distribution. In the<br />

work at hand, we present an integrated approach which<br />

addresses the Service Selection Problem for complex workflows<br />

in conjunction with stochastic Quality of Service parameters.<br />

Accounting for penalty cost which accrue due to Quality<br />

of Service violations, our approach reduces the impact of<br />

stochastic QoS behavior on total cost significantly.<br />

Keywords-Service Selection, Stochastic Quality of Service,<br />

Optimization, Simulation<br />

I. INTRODUCTION<br />

The selection of services from a set of appropriate ones<br />

that are able to provide the required functionality and thereby<br />

best meeting cost and Quality of Service (QoS) requirements<br />

– the Service Selection Problem (SSP) – is widely<br />

recognized in the literature, e.g., [1]–[4]. The optimization<br />

of the SSP is thereby based on deterministic QoS values.<br />

A solution to the SSP describes an execution plan, i.e., an<br />

assignment of services to certain tasks of a workflow, which<br />

satisfy the mentioned cost and QoS constraints.<br />

But QoS, e.g., the response time of a service or its<br />

availability, is not always deterministic in reality. Due to<br />

network latency or server load, response times of services<br />

may change dynamically. I.e., when the execution of the<br />

computed execution plan actually takes place, the perceived<br />

QoS might differ from the expected QoS which has previously<br />

been used for the calculation of the execution plan.<br />

Thus, although having computed an optimal solution to the<br />

SSP during design time which satisfies the constraints, it still<br />

is possible that these constraints are violated during runtime.<br />

The work at hand addresses this issue. Based on a service<br />

broker scenario, which is presented in Section II, we describe<br />

how QoS violations due to stochastic QoS behavior negatively<br />

impacts total cost. In order to account for this impact<br />

of stochastic QoS parameters, we present an integrated<br />

approach comprising an optimization, a simulation, and an<br />

adaptation step.<br />

Stefan Schulte<br />

Distributed Systems Group<br />

Vienna University of Technology<br />

Vienna, Austria<br />

Email: s.schulte@infosys.tuwien.ac.at<br />

During the optimization step, we compute an optimal<br />

solution to the SSP, i.e., an optimal execution plan, satisfying<br />

the QoS constraints based on the deterministic QoS values<br />

denoted by the respective service providers. In the simulation<br />

step, we observe the expected runtime behavior of the<br />

computed execution plan in terms of QoS. This way, we can<br />

assess potentially occurring constraint violations. We thereby<br />

assume that violating QoS constraints is penalized, i.e.,<br />

penalty fees become due in addition to service invocation<br />

cost. According to the results of the simulation, we apply a<br />

greedy adaptation heuristic in order to reduce the impact of<br />

potentially occurring constraint violations.<br />

The remainder of this work is structured as follows. In<br />

Section II, we present a motivating scenario, which will be<br />

used throughout the paper. In Section III, we describe our<br />

solution to the SSP, which is based on our previous work<br />

in [5]. The applied simulation approach is presented in Section<br />

IV. Based on the simulation results, we apply our greedy<br />

adaptation heuristic which is described in Section V and<br />

evaluated in Section VI. Finally, after having distinguished<br />

our approach from related work in Section VII, we draw<br />

conclusions and discuss future work in Section VIII.<br />

II. SCENARIO<br />

In this section, we present a scenario, which is used as<br />

an example in the work at hand in order to illustrate the<br />

impact of stochastic QoS parameters. The application of our<br />

approach is not limited to this scenario.<br />

Imagine a service broker who receives requests from its<br />

customers. Paying the broker a fixed amount of money,<br />

the customers require certain tasks and workflows, respectively,<br />

to be executed. For this, they provide the broker<br />

with a document which specifies the required tasks from<br />

a functional perspective and indicates the ordering of the<br />

tasks, i.e., the structure of the workflow. One of the broker’s<br />

customers asks for instance for the workflow in Figure 1.<br />

The process steps P Si thereby indicate the tasks which have<br />

to be accomplished. Each task can be executed by a single<br />

service.


PS1<br />

0.7<br />

0.3<br />

X<br />

PS5<br />

PS3<br />

PS2<br />

PS7<br />

Figure 1: Example Workflow (in BPMN)<br />

In addition to these functional requirements, the customers<br />

also specify their QoS needs regarding the execution of the<br />

respective workflow. For this, they provide restrictions in the<br />

form of upper or lower bounds for specific QoS parameters,<br />

the so-called Service Level Objectives (SLOs). With this<br />

information, the broker tries to select those services among<br />

functionally appropriate ones which satisfy the customers’<br />

QoS requirements, as a violation of these requirements will<br />

be penalized by the broker’s customers. Having selected respective<br />

services, the broker pays and invokes these services<br />

in order to execute the customers’ tasks and workflows, respectively.<br />

If the customers’ QoS requirements are violated,<br />

penalty fees become due, which also have to be paid by<br />

the broker. In order to reach an optimized decision, the<br />

broker models the selection of services as an optimization<br />

problem aiming at minimizing invocation cost and satisfying<br />

QoS constraints. I.e., the broker formulates an SSP, which<br />

is described in the following Section III.<br />

PS4<br />

PS6<br />

III. SERVICE SELECTION PROBLEM<br />

In order to formulate an SSP and therewith to compute<br />

an execution plan, it is necessary to aggregate the QoS and<br />

cost values of eligible candidate services according to the<br />

regarded workflow structures. For this, we specify a system<br />

model in Section III-A and discuss respective aggregation<br />

functions in Section III-B. Finally, we utilize the presented<br />

aggregation functions and provide an optimization model in<br />

Section III-C.<br />

A. System Model<br />

In this section, we describe the system model utilized in<br />

the work at hand. The set of all tasks of a workflow is<br />

labeled with I, i ∈ I = {1, ..., i # }. The set of services<br />

appropriate to accomplish a certain task i is labeled with Ji,<br />

j ∈ Ji = {1, ..., j #<br />

i }. The decision variables xij ∈ {0, 1}<br />

indicate whether a service j is selected to accomplish task i.<br />

According to our running example, we consider cost c<br />

(charge for invoking a service in cent applying a pay-peruse<br />

pricing model), response time r (time elapsed between<br />

invoking a service and receiving its response), availability<br />

a (probability that a service is available), and throughput d<br />

(number of requests a service is able to serve within a certain<br />

time interval). These parameters – in fact, even a subset of<br />

these parameters – are sufficient to cover the aggregation<br />

types summation, multiplication, and min/max-operator (cf.<br />

X<br />

PS8<br />

Section III-B). Further QoS parameters can be integrated<br />

into the optimization problem straightforwardly. Bounds for<br />

these parameters are labeled with bc, br, ba, bd.<br />

Regarding branchings, we label the set of paths with<br />

L, in which l ∈ L = {1, ..., l # } indicates the respective<br />

path number. Referring to the workflow in Figure 1, there<br />

are two paths l within the AND-block, thus L = {1, 2}.<br />

Different sets of paths will be distinguished by utilizing<br />

additional indices, i.e., La, Lx for AND/XOR. We refer to<br />

them as branching La, Lx. The tasks within a branching are<br />

covered by the set IL ⊆ I, whereas Il ⊆ IL represents<br />

the set of tasks within path l. We label the set of the<br />

remaining tasks, which are not located within a branching,<br />

with Is = I \ (Il | l ∈ L). Utilizing this system model, we<br />

develop aggregation functions in the following.<br />

B. Aggregation Functions<br />

As previously stated, it is necessary to aggregate the<br />

QoS and cost values of candidate services according to<br />

regarded workflow patterns in order to compute the overall<br />

cost and QoS for a specific workflow. Regarding our example<br />

scenario, this actually is a prerequisite for comparing<br />

workflow QoS with respective bounds issued by the broker’s<br />

customers. As previously stated, violation of the customers’<br />

bounds will result in additional penalty cost. Thus, the<br />

broker has a high interest in making sure that the customers’<br />

bounds are satisfied. For this, the broker performs a worstcase<br />

analysis as opposed to a best-case or average-case<br />

analysis. While probabilities for the execution of certain<br />

paths of a branching would be considered in an averagecase<br />

analysis, the worst (best) paths of each branching in<br />

terms of aggregated cost and QoS are considered in a worstcase<br />

(best-case) analysis for the computation of an optimal<br />

solution to the SSP. Respective aggregation functions for<br />

sequences, AND-blocks, and XOR-blocks are indicated in<br />

Table I.<br />

Table I: Worst-Case Aggregation Functions<br />

Sequence AND-block XOR-block<br />

� �<br />

� �<br />

� �<br />

rijxij max rijxij max rijxij<br />

l∈L<br />

l∈L<br />

i∈Is � j∈J �i<br />

� �i∈I<br />

l j∈J � i<br />

i∈I �l<br />

j∈J �i<br />

aijxij<br />

aijxij min aijxij<br />

l∈L<br />

i∈Is j∈J �i<br />

l∈L i∈Il j∈Ji i∈Il j∈Ji min dijxij min<br />

i∈Is<br />

l∈L<br />

j∈Ji (min<br />

�<br />

dijxij) min<br />

i∈Il l∈L<br />

j∈Ji (min<br />

�<br />

dijxij)<br />

i∈Il � � � � �<br />

� j∈J �i<br />

cijxij<br />

cijxij max cijxij<br />

l∈L<br />

i∈Is j∈Ji l∈L i∈Il j∈Ji i∈Il j∈Ji For a sequence, the cost and QoS values of all services<br />

selected to accomplish certain tasks i have to be aggregated<br />

according to the respective aggregation type, e.g., summed<br />

up for cost. Regarding AND-blocks, it has to be noted that<br />

for response time r only the path with the highest aggregated<br />

response time – the critical path – requires consideration, as<br />

the tasks within the different paths are executed in parallel.


Regarding the other non-functional parameters, the cost<br />

and QoS values of all services have to be aggregated as<br />

all selected services are executed in the end (similar to a<br />

sequence). For XOR-blocks, where only one of the potential<br />

paths is executed, we consider the path with the worst<br />

aggregated cost and QoS values according the respective<br />

aggregation types as we pursue a worst-case analysis.<br />

In the work at hand, we stick to these patterns for the<br />

sake of simplicity. The interested reader may refer to our<br />

former work in [5] for further structured workflow patterns<br />

(OR-blocks, Repeat Loops) as well as for unstructured patterns<br />

of Directed Acyclic Graphs and respective aggregation<br />

functions, which could have additionally been utilized.<br />

Applying the presented aggregation functions, we formulate<br />

the SSP as an optimization problem in the following.<br />

C. Optimization Model<br />

In this section, we formulate the SSP in Model 1 –<br />

accounting for the example workflow in Figure 1. For<br />

this, we specify an objective function in (1) and a set of<br />

restrictions in (2)–(10) by applying the aggregation functions<br />

from Table I.<br />

Model 1 Service Selection Problem for Example Workflow<br />

Objective Function<br />

minimize � �<br />

cijxij + �<br />

(c ′ l + � �<br />

cijxij) (1)<br />

so that<br />

i∈Is j∈Ji<br />

i∈Is j∈Ji<br />

l∈La<br />

l∈La<br />

i∈Il j∈Ji<br />

� �<br />

rijxij + max(r<br />

l∈La<br />

i∈Is j∈Ji<br />

′ l + � �<br />

rijxij) ≤ br (2)<br />

i∈Il j∈Ji<br />

( � �<br />

aijxij) · ( �<br />

(a ′ l · � �<br />

aijxij)) ≥ ba (3)<br />

min(min<br />

i∈Is<br />

�<br />

i∈Il j∈Ji<br />

dijxij, min(d<br />

l∈La<br />

j∈Ji<br />

′ l, min<br />

i∈Il<br />

j∈Ji<br />

�<br />

dijxij))) ≥ bd (4)<br />

max (<br />

lx∈Lx<br />

� �<br />

rijxij) = r<br />

i∈Ilx j∈Ji<br />

′ l ∀l ∈ La| interlaced (5)<br />

min (<br />

lx∈Lx<br />

� �<br />

aijxij) = a<br />

i∈Ilx j∈Ji<br />

′ l ∀l ∈ La| interlaced (6)<br />

max<br />

lx∈Lx<br />

(min(<br />

�<br />

dijxij)) = d ′ l ∀l ∈ La| interlaced (7)<br />

i∈Ilx<br />

j∈Ji<br />

max (<br />

lx∈Lx<br />

� �<br />

cijxij) = c<br />

i∈Ilx j∈Ji<br />

′ l ∀l ∈ La| interlaced (8)<br />

�<br />

xij = 1 ∀i ∈ I (9)<br />

j∈Ji<br />

xij ∈ {0, 1} ∀i ∈ I, ∀j ∈ Ji (10)<br />

Note that the considered workflow in Figure 1 contains<br />

an XOR-block within the AND-block. As the aggregation<br />

functions indicated in Table I assume the tasks within splits<br />

and corresponding joins to be arranged sequentially (cf. [6]),<br />

we have to account for this interlaced structure. Referring to<br />

our former work in [6], we abstract from the interlacing and<br />

insert additional variables r ′ l , c′ l , a′ l , d′ l<br />

and their respective<br />

aggregation functions (i.e., for XOR-blocks in this case)<br />

into Model 1. As the broker receives a fixed amount of<br />

money from his/her customers for satisfying their needs,<br />

(s)he maximizes his/her profit by selecting and invoking<br />

the services with minimal cost. Thus, the objective function<br />

in (1) aims at minimizing service invocation cost. The<br />

customers’ restrictions on cost and QoS are indicated in (2)–<br />

(7). In (9), we make sure that only one service is selected<br />

for each task, and (10) represents the integrality condition<br />

for the decision variables.<br />

Model 1 constitutes a non-linear optimization problem as<br />

it contains non-linear aggregations of decision variables xij,<br />

i.e., multiplication and min/max-operator in (2), (3), (4) for<br />

instance. We transform it into a linear one by linearizing the<br />

non-linear restrictions in (3)–(7). Due to space restrictions,<br />

we omit describing the linearization in the work at hand<br />

and refer the interested reader to our former work in [5],<br />

[6]. The optimal solution to the obtained linear optimization<br />

problem can then be computed by applying Integer Linear<br />

Programming (ILP) techniques from the field of Operations<br />

Research [7].<br />

Thus, having computed an optimal execution plan minimizing<br />

service invocation cost and (presumably) satisfying<br />

the customers’ constraints, the broker would apply this<br />

solution and invoke the respective services while assuming<br />

that they actually hold the constraints. But as the invoked<br />

services may show a different behavior during runtime than<br />

expected during design time, the application of the computed<br />

execution plan could lead to violations of the customers’<br />

constraints. This would result in additional cost for the<br />

broker as, in this case, penalty fees will become due.<br />

In order to assess the impact of stochastic QoS values,<br />

we propose to perform an additional simulation step, which<br />

is described in the following Section IV.<br />

A. General Concept<br />

IV. SIMULATION<br />

The simulation approach presented in the work at hand<br />

is substantially inspired by findings from the domain of<br />

project management. In our former work in [8], we outlined<br />

the conceptual similarities between workflows and project<br />

networks, specifically, generalized activity networks [9]: a<br />

complex workflow can easily be interpreted as a project<br />

network, which consists of comparable entities, such as tasks<br />

and branches.<br />

Based on the work in [10], we argued in [8] that simulation<br />

provides the best means to assess the risk of breaking


given QoS constraints in practical application. Referring to<br />

[10], deterministic methods regularly fail to correctly quantify<br />

such risks, specifically if large and complex networks<br />

and workflows, respectively, are concerned.<br />

Thus, regarding our example scenario, an optimization<br />

approach based on deterministic QoS values does not capture<br />

the risk of violating QoS constraints. For instance, if a<br />

selected service experiences unusually high demand, it may<br />

not be able to provide a certain response time, even if a<br />

corresponding bound has been guaranteed by the service<br />

provider. Thus, in turn, an execution plan may not satisfy<br />

the QoS requirements in some cases. The existence of<br />

substantial QoS fluctuations – specifically with respect to<br />

Web service response times – has been empirically shown,<br />

for instance by Rosario et al. [11] and Miede et al. [12]. In<br />

the work at hand, we quantify these inherent risks through<br />

simulation, i.e., repeated “virtual” execution of a workflow.<br />

B. Stochastic QoS Parameters<br />

As previously stated, it is assumed in most related approaches<br />

addressing the SSP that a “hard” – deterministic<br />

– QoS guarantee is specified for each service candidate and<br />

QoS parameter. In order to perform a simulation, we further<br />

assume that probabilistic QoS specifications are available.<br />

In accordance with the notation in Section III-A, Rij, Aij,<br />

and Dij represent random variables for the QoS parameters<br />

response time, availability, and throughput of service j<br />

for task i. These random variables may follow arbitrary<br />

probability distributions. For instance, the response time of<br />

a service may be modeled using a normal distribution, i.e.,<br />

Rij ∼ N(100, 20). For an overview of common probability<br />

distributions, we refer to corresponding compendiums [13].<br />

Probability distributions can be deduced in at least two<br />

principle ways: First, they may be explicitly provided by a<br />

service provider as part of contracted guarantees in terms of<br />

Service Level Agreements (SLA), following the idea of “soft<br />

contracts” [11]. Second, they may be inferred by mining the<br />

monitoring data from past service executions.<br />

C. Execution of the Simulation<br />

For the actual simulation, we virtually execute the previously<br />

computed execution plan a predefined number of<br />

times. In each iteration, we draw a specific realization of<br />

each QoS parameter for each selected service, based on<br />

the given random variables Rij, Aij, and Dij. In addition,<br />

depending on the branching probabilities of XOR-splits, we<br />

draw random variables in order to determine which paths<br />

are actually executed in the current iteration. The realized<br />

values of the individual services are aggregated afterwards<br />

according to the respective workflow structure applying the<br />

aggregation functions from Table I. This way, we compute<br />

the overall QoS values for the whole workflow.<br />

Based on the workflow in Figure 1, we provide an<br />

example in Figure 2. It represents an execution plan where<br />

r11 = 100<br />

R11 ~ N(95, 5)<br />

S11<br />

br = 430<br />

0.7<br />

0.3<br />

S51<br />

S21<br />

X X<br />

S31<br />

r31 = 116<br />

R31 ~ N(110, 7)<br />

r51 = 111<br />

R51 ~ N(100, 12)<br />

r21 = 221<br />

R21 ~ N(220, 23)<br />

S71<br />

r71 = 224<br />

R71 ~ N(210, 15)<br />

S41<br />

r41 = 107<br />

R41 ~ N(95, 13)<br />

S61<br />

r61 = 112<br />

R61 ~ N(105, 8)<br />

r81 = 106<br />

R81 ~ N(100, 7)<br />

Figure 2: Workflow execution plan with services’ QoS<br />

guarantees, serving as input to the simulation.<br />

possible, appropriate services have been identified for the<br />

realization of all tasks. Each service is associated with<br />

deterministic QoS bounds, as specified by respective service<br />

providers (boxes with light-gray background). Further,<br />

probabilistic QoS specifications – as observed by the broker<br />

– are indicated (boxes with dark-gray background). Note<br />

that we assume conservative service providers as the deterministic<br />

QoS values specified by the service providers are<br />

higher than the expected values. For simplicity reasons, we<br />

only incorporate the QoS parameter response time in this<br />

example. As can be quickly validated, the workflow will<br />

meet the QoS constraint, namely br = 430, according to the<br />

given deterministic QoS values.<br />

However, performing a simulation with 10, 000 iterations,<br />

i.e., drawing 10, 000 corresponding realizations of<br />

the regarded QoS parameter response time, reveals that the<br />

specified QoS constraint will be violated in approximately<br />

15% of all executions – although we assumed conservative<br />

service providers. Figure 3 indicates the respective aggregated<br />

response time for the workflow. The QoS violation<br />

can be attributed to one or more services exceeding their<br />

deterministic QoS guarantees, which cannot be sufficiently<br />

captured by the initial service selection and, thus, will result<br />

in potentially severe penalities for the broker. In order to<br />

reduce the impact of stochastic QoS behavior and therewith<br />

the accruing penalties, the boker applies our greedy adaptation<br />

heutistic, which is discribed in the following Section V.<br />

V. GREEDY ADAPTATION HEURISTIC<br />

In the previous section, we have outlined how a simulation<br />

step may reveal the risk of violating certain QoS constraints<br />

in practice. This knowledge can be exploited in order to minimize<br />

the risk of penalties and therewith the total cost for the<br />

broker in our example scenario comprising cost for invoking<br />

the selected services and cost according to expected penalty<br />

fees. For this, we present a greedy adaptation heuristic in<br />

this section aiming at reducing the total cost for the broker.<br />

In this context, adaptation denotes excluding and replacing,<br />

respectively, those services from the formerly computed<br />

execution plan for other functionally appropriate services,<br />

S81


Cumulative probability<br />

1<br />

0.9<br />

0.8<br />

0.7<br />

0.6<br />

0.5<br />

0.4<br />

0.3<br />

0.2<br />

0.1<br />

0<br />

350 360 370 380 390 400 410 420 430 440 450 460 470<br />

Workflow response time<br />

Figure 3: Cumulative probability function of the workflow’s<br />

response time, according to the simulation.<br />

which cause high penalty cost due to unexpected runtime behavior.<br />

One may argue that we could have also performed an<br />

adaptation by trying to improve QoS for the services of the<br />

computed execution plan, i.e., asking the respective service<br />

providers to either enhance resource deployment in terms<br />

of, e.g., main storage and CPU, or to (generally) improve<br />

the implementation of the service providers’ services. As<br />

both is out of the broker’s control sphere, we do not focus<br />

on nor account for such possibilities. We further assume<br />

SLAs to be fixed and do not account for SLA negotiations,<br />

neither between the broker and its customers, nor between<br />

the broker and respective service providers – which could<br />

also be seen as an alternative adaptation in order to tackle<br />

QoS violations – as (optimal and efficient) SLA negotiation<br />

is a research topic on its own and would go beyond the<br />

scope of the paper at hand.<br />

Regarding penalty cost, we assume linear penalty fees per<br />

unit of QoS violation, e.g., cents per second the execution<br />

took longer than restricted by the respective bound, or cent<br />

per percent point the availability was lower than restricted.<br />

It would also be imaginable to assume variable penalty fees,<br />

which increase quadratically or exponentially with the size<br />

of the violation. As this would only influence the calculation<br />

of the actual total penalty cost depending on the chosen<br />

penalty model but does not change our approach, we stick<br />

to linear penalty fees for the sake of simplicity.<br />

Our greedy adaptation heuristic is indicated in Algorithm<br />

1, using pseudocode. The heuristic is split into four<br />

steps. First, we determine the total cost of the current solution.<br />

Second, we compute the critical QoS parameter, i.e.,<br />

the QoS parameter which causes the highest penalty cost, as<br />

this parameter might bear the highest penalty cost savings.<br />

Third, we determine for which of the tasks i ∈ I a potential<br />

improvement of the respective critical QoS parameter would<br />

be highest. Fourth, we finally perform the adaptation.<br />

More details on these steps are provided in the following.<br />

Referring to Algorithm 1, we compute an optimal solu-<br />

tion cs, which represents the current solution, based on<br />

deterministic QoS values, perform the previously mentioned<br />

simulation step, and calculate the QoS violation in lines<br />

2 to 4. In order to determine the total cost of the current<br />

solution, we aggregate the invocation and accruing penalty<br />

cost as indicated in line 5. By computing and comparing the<br />

penalty cost for each QoS parameter, we determine critical<br />

QoS parameter in lines 7 to 13. In order to compute the<br />

critical task, we compute in line 16 the potential benefit for<br />

each task i. In this context, the potential benefit of a task<br />

i corresponds to the highest possible reduction in standard<br />

deviation σ. Referring to (11), we compute the difference<br />

of the selected service’s σs to the σj values of the other<br />

candidate services of task i.<br />

benefit = max(σs<br />

− σj) (11)<br />

j∈Ji<br />

The highest possible reduction in σ, which corresponds<br />

to the highest possible reduction in uncertainty and risk<br />

regarding stochastic QoS parameters, thereby represents the<br />

potential (absolute) benefit for that task. In line 17, we<br />

compute the weight ω for task i. For this, we divide the<br />

number of the simulation runs where task i has been virtually<br />

executed and the restriction for the critical QoS parameter<br />

has been violated by the total number of runs. As indicated<br />

in lines 18 to 20, the task with the highest weighted benefit<br />

becomes the critical task, which is required in order to<br />

perform the actual adaptation.<br />

As stated earlier, our adaptation comprises excluding and<br />

replacing services while other adaptation techniques and<br />

mechanisms also would have been possible and supported by<br />

our heuristic. Insofar, our approach is flexible and extensible<br />

for supporting further adaptation techniques. We determine<br />

the currently selected service js in line 24. We ban in lines<br />

25 to 27 all services of the critical task, which have a<br />

negative benefit, i.e., whose σ and therewith the risk of QoS<br />

violation is larger than or equal to the σjs<br />

of the currently<br />

selected service js. This way, we adapt the list of available<br />

candidate services for the critical task.<br />

Utilizing this adapted list, we rerun the optimization and<br />

obtain a new execution plan, which is optimal under the<br />

adapted circumstances. Afterwards, we conduct the simulation<br />

step again and calculate both the invocation and penalty<br />

cost for the lastly computed execution plan (cf. the first<br />

step). By comparing the total cost of this new solution with<br />

the previous – formerly known as current best – solution,<br />

we determine whether applying our adaptation heuristic has<br />

been advantageous from the broker’s point of view. Via a<br />

parameter greed, we specify the algorithm’s degree of greed,<br />

i.e., whether and how often the described steps are repeated<br />

as long as past iterations reduced total cost. Via a further<br />

parameter anneal, we control for allowing worse solutions<br />

temporarily as starting point for a continuous application of<br />

the algorithm. An evaluation for this approach is presented<br />

in the following Section VI.


Algorithm 1 Greedy Adaptation Heuristic<br />

1: //First step – determine current cost<br />

2: cs = computeCurrentSolution();<br />

3: sim = simulate(cs);<br />

4: v = computeQoSViolation(sim);<br />

5: totalCost = computeInvCost(cs) + computePanalty(v);<br />

6: //Second step – determine the critical QoS parameter<br />

7: for all p ∈ QoSparameters do<br />

8: penaltyCost = computePenalty(p);<br />

9: if penaltyCost ≥ highestP enaltyCost then<br />

10: highestP enaltyCost = penaltyCost;<br />

11: criticalQoSparameter = p;<br />

12: end if<br />

13: end for<br />

14: //Third step – determine the critical task<br />

15: for all i ∈ I do<br />

16: benefit = computeBenefit(i);<br />

17: ω = computeWeight(i);<br />

18: if ω · benefit ≥ highestBenefit then<br />

19: highestBenefit = ω · benefit;<br />

20: criticalT ask = i;<br />

21: end if<br />

22: end for<br />

23: //Fourth step – perform the adaptation<br />

24: js = getSelectedServiceOf(i);<br />

25: for all j ∈ Ji do<br />

26: if σj ≥ σjs then<br />

27: setBanned(j);<br />

28: end if<br />

29: end for<br />

VI. EVALUATION<br />

As a proof of concept, we implemented our greedy adaptation<br />

heuristic. For the computation of an optimal solution to<br />

the SSP based on deterministic values, we utilized the linear<br />

programming solver CPLEX 1 . In this section, we evaluate<br />

the impact of QoS violation on total cost with respect to our<br />

broker scenario. For this, we conducted a set of experiments<br />

in order to assess the effects of different configurations<br />

regarding the adaptation parameters greed and anneal, which<br />

allow to control for the number of iterations the algorithm<br />

performs in order to achieve improved solutions and for the<br />

number of thereby temporarily accepted worse solutions.<br />

In the following, we describe our experimentation setup.<br />

We consider the workflow in Figure 1, which contains P S1-<br />

P S8 indicating tasks i ∈ {1, ..., 8}. In order to determine<br />

QoS values of the respective services, we draw realizations<br />

of the random variables Rij, Aij, and Dij. We assume<br />

these random variables to follow a normal distribution, i.e.,<br />

Rij ∼ N(µr, σr), Aij ∼ N(µa, σa), Dij ∼ N(µd, σd).<br />

1 http://www.ibm.com/software/integration/optimization/cplex-optimizer/<br />

Table II: Stochastic QoS<br />

QoS P S2, P S7 P S1, P S3 − P S6, P S8<br />

Rij<br />

Aij<br />

Dij<br />

cij<br />

µr ∼ U(160, 240)<br />

σr ∼ U(0, 40)<br />

µa ∼ U(0.92, 1.0)<br />

σa ∼ U(0.0, 0.08)<br />

µd ∼ U(80, 120)<br />

σd ∼ U(0, 30)<br />

U(0.8, 1.2) · (40+<br />

(0.03 · (µd − µr)) · µ 2 a )<br />

µr ∼ U(80, 120)<br />

σr ∼ U(0, 20)<br />

µa ∼ U(0.94, 0.98)<br />

σa ∼ U(0.0, 0.04)<br />

µd ∼ U(80, 120)<br />

σd ∼ U(0, 30)<br />

U(0.8, 1.2) · (20+<br />

(0.03 · (µd − µr)) · µ 2 a )<br />

We could have equally utilized other distribution functions<br />

or inferred the respective distributions by mining the<br />

monitoring data from past service executions, which we<br />

actually envisage in our future work, but we stick to normal<br />

distributions in the work at hand for the sake of simplicity.<br />

The parameterization of the random variables Rij, Aij, and<br />

Dij is indicated in Table II. We assume that the invocation<br />

cost of a service partly depends on its QoS, i.e., good QoS<br />

values in terms of low response time r, high availability a,<br />

and high throughput d result in higher invocation cost. For<br />

this, as also indicated in Table II, we compute the invocation<br />

cost of a service according to its QoS, utilizing an additional,<br />

uniform distributed random variable U(a, b).<br />

In order to assess the impact of the number of beneficial<br />

iteration steps – as indicated by the parameter greed –<br />

on total cost and computation time, we varied greed in<br />

Figure 4a and Figure 4d from 0 to 20 step two, utilizing<br />

anneal = 4. Regarding the influence of the number of<br />

temporarily allowed worse solutions, we varied the anneal<br />

parameter in Figure 4b from 0 to 16 step two, utilizing a<br />

fixed greed value of 10. For these experiments, we set the<br />

violation cost to 10% of the respective service invocation<br />

cost – per unit of QoS violation (cf. Section V). In Figure 4c,<br />

we account for different penalty cost by varying the penalty<br />

cost percentage from 0% to 20%, utilizing greed = 10,<br />

anneal = 4. Finaly, in Figure 4d, the computation time for<br />

computing respective solutions is indicated. The experiments<br />

were performed on an Intel Core 2 Quad processor at<br />

2.66 GHz, 4 GB RAM, running Microsoft Windows 7.<br />

The evaluation results show that the application of our<br />

greedy adaptation heuristic to the considered service broker<br />

scenario leads to a cost reduction of 9 ct to 12.5 ct in relation<br />

to total cost of 142.5 ct for the whole workflow, which corresponds<br />

to a reduction of 6% to 8.5%. Thus, the broker could<br />

save 6% to 8.5% of the total cost. But, Figure 4d reveals<br />

that reducing the cost actually “costs” computation time –<br />

up to 10 times as much. As Figure 4a indicates, additional<br />

reduction in total cost decreases with additional adapatation<br />

steps, i.e., higher greed values. Therefore, utilizing a greed<br />

value of 4 to 6, for which the computation time is roughly<br />

6 times higher, seems to be a good compromise between<br />

cost reduction and additional computation time. Also values<br />

greater than 4 for anneal do not improve the cost reduction


Total Cost (in ct)<br />

165<br />

160<br />

155<br />

150<br />

145<br />

140<br />

135<br />

130<br />

No adapatation<br />

Adaptation heuristic<br />

125<br />

0 5 10<br />

Greed<br />

15 20<br />

(a) Impact of greed<br />

Total Cost (in ct)<br />

165<br />

160<br />

155<br />

150<br />

145<br />

140<br />

135<br />

130<br />

No adapatation<br />

Adaptation heuristic<br />

125<br />

0 2 4 6 8 10 12 14 16<br />

Annealing<br />

(b) Impact of anneal<br />

significantly. Thus, using the parameterization greed = 6,<br />

anneal = 4, which leads to a cost reduction of 7.3% and a<br />

6-times magnified computation time, appears sensible.<br />

Having described our approach, we discuss related approaches<br />

in the following Section VII.<br />

VII. RELATED WORK<br />

As previously stated, the SSP is widely recognized in<br />

the literature. A survey of current approaches can be found<br />

in [4]. In principle, current approaches can be divided into<br />

two categories: heuristic approaches which try to find rather<br />

good solutions within a reduced amount of computation<br />

time, e.g., [2], [14], [15], and approaches aiming at finding<br />

an optimal solution to the SSP, e.g., [1], [3], [16]. All those<br />

approaches presume that the utilized QoS parameters are<br />

deterministic. Relevant related work in the area of stochastic<br />

QoS parameter, however, is rather sparse.<br />

In their work in [11], Rosario et al. consider probabilistic<br />

QoS values, but not for the purpose of services selection.<br />

They rather focus on SLA and contract composition, respectively,<br />

using soft probabilistic contracts as already stated in<br />

Section IV-A. The authors in [17] pay insofar attention to<br />

stochastic QoS parameters as they try to achieve an accurate<br />

prediction of QoS values based on historic data which<br />

then is utilized for service selection rather than predefined<br />

values guaranteed by service providers. In [18], Hwang<br />

et al. utilize Probability Mass Functions (PMFs) for QoS<br />

instead of deterministic values. They describe approaches for<br />

aggregating the PMFs of single services. The authors thereby<br />

utilize a preselected set of services with discrete PMFs, i.e.,<br />

they do not perform service selection based on PMFs, but<br />

rather aim at computing and estimating QoS for servicebased<br />

workflows. In contrast to this, our approach targets<br />

service selection accounting for stochastic QoS parameters.<br />

Cardellini et al. consider stochastic QoS parameters for the<br />

SSP insofar as they use an α-percentile (with α = 95%)<br />

for the QoS parameter response time [19]. Accordingly,<br />

instead of utilizing deterministic response time values for the<br />

optimization, the authors integrate a restriction demanding<br />

the probability of violating the bound for response time to<br />

be lower or equal to 1 − α, i.e., 1 − 0.95 = 5%. Projected to<br />

our broker scenario, this means that the broker can assume<br />

Total Cost (in ct)<br />

165<br />

160<br />

155<br />

150<br />

145<br />

140<br />

135<br />

130<br />

Figure 4: Evaluation Results<br />

No adapatation<br />

Adaptation heuristic<br />

125<br />

0 0.05 0.1<br />

Penalty Fee<br />

0.15 0.2<br />

(c) Impact of penalty cost<br />

Computation Time (in msec)<br />

900<br />

800<br />

700<br />

600<br />

500<br />

400<br />

300<br />

200<br />

100<br />

0<br />

No adapatation<br />

Adaptation heuristic<br />

0 5 10<br />

Greed<br />

15 20<br />

(d) Computation time<br />

satisfying the respective bound with a probability of 95%.<br />

But Cardellini et al. thereby fail to account for the arising<br />

penalty cost due to QoS violations in the remaining 5% of<br />

the cases. Our approach, on the other hand, proceeds one<br />

step beyond and considers the impact of QoS violations in<br />

terms of penalty cost. Depending on the ratio between invocation<br />

and penalty cost, it could be beneficial for the broker<br />

selecting rather cheap services, which statistically cause QoS<br />

violations more often, and paying the penalty cost rather<br />

than selecting expensive services with low probabilities of<br />

violating QoS constraints. Thus, our approach enables the<br />

broker to select those services that bear the lowest cost.<br />

In the approach in [20], which probably comes closest to<br />

ours, Leitner et al. assume a fixed service composition and<br />

a fixed set of possible adaptations to improve the service<br />

composition in terms of, e.g., “utilize Express Shipping”<br />

instead of “ utilize Standard Shipping”. The aim is to select<br />

and apply those adaptations that minimize cost comprising<br />

invocation cost, penalty cost for QoS violation, and cost for<br />

applied adaptations, which aim at avoiding QoS violation. If<br />

we abstract from the term adaptation and interpret available<br />

adaptations as alternative services, then Leitner et al. are<br />

solving a SSP with the aim of minimizing total cost, at which<br />

penalty cost for QoS violation are considered as well. In<br />

order to account for non-deterministic QoS behavior during<br />

runtime, the authors utilize a predictor component which<br />

predicts prospective QoS values and therewith expected QoS<br />

violation. Thus, Leitner et al. estimate the impact of stochastic<br />

QoS behavior for each service separately and perform a<br />

optimization with these estimated, deterministic QoS values<br />

allowing for QoS violations and accounting for their impact<br />

on total cost. Our approach, however, goes one step further,<br />

as we do not consider the stochastic QoS behavior of the<br />

services independently of each other, but account for the<br />

whole workflow during our simulation step. Thus, potential<br />

reverse QoS deviations of different services from expected<br />

behavior can be considered, which is not possible in [20] due<br />

to their isolated consideration of expected QoS per service,<br />

independently of other services.<br />

In summary, our approach extends related work as it<br />

considers, on the one hand, the impact of QoS violation in<br />

terms of accruing penalty costs. On the other hand, we do not


only regard isolated stochastic QoS behavior for individual<br />

services, but account for probably compensating reverse<br />

QoS deviations of different services. Thus, we consider the<br />

impact of stochastic QoS behavior for the whole workflow.<br />

Conclusions are drawn in the following Section VIII.<br />

VIII. CONCLUSION<br />

The SSP is widely recognized in the literature and has<br />

been discussed in several scientific papers – based on deterministic<br />

QoS parameters. In the work at hand, we addressed<br />

the SSP in conjunction with stochastic QoS parameters<br />

which has been considered as yet only insufficiently in the<br />

literature. For this, we presented an integrated approach<br />

comprising an optimization, a simulation, and an adaptation<br />

step which aims at reducing the impact of stochastic QoS<br />

behavior on total cost. The evaluation shows that the application<br />

of our approach leads to a cost reduction up to 8.5%,<br />

utilizing the described service broker scenario. Thus, the<br />

actual, absolute level of cost reduction depends on the concrete<br />

paramerization and the regarded scenario. For this, we<br />

will extend the evaluation in our future work by considering<br />

further workflow structures and QoS distribution functions as<br />

well as different degrees of conservative, deterministic QoS<br />

values issued by service providers. In addition, we focus on<br />

improving the scalability of our greedy adaptation heuristic.<br />

ACKNOWLEDGMENT<br />

This work is supported in part by the Commission of the<br />

European Union within the ADVENTURE FP7-ICT project<br />

(Grant agreement no. 285220) and by E-<strong>Finance</strong> <strong>Lab</strong> e. V.,<br />

Frankfurt am Main, Germany (http://www.efinancelab.com).<br />

REFERENCES<br />

[1] D. Ardagna and B. Pernici, “Adaptive service composition in<br />

flexible processes,” IEEE Transactions on Software Engineering<br />

(TSE), vol. 33, no. 6, pp. 369–384, 2007.<br />

[2] D. A. Menascé, E. Casalicchio, and V. K. Dubey, “A Heuristic<br />

Approach to Optimal Service Selection in Service Oriented<br />

Architectures,” in Workshop on Software and Performance<br />

(WOSP). ACM, 2008, pp. 13–24.<br />

[3] A. F. M. Huang, C.-W. Lan, and S. J. H. Yang, “An Optimal<br />

QoS-based Web Service Selection Scheme,” Information<br />

Sciences (ISCI), vol. 179, no. 19, pp. 3309–3322, 2009.<br />

[4] A. Strunk, “QoS-Aware Service Composition: A Survey,” in<br />

European Conf. on Web Services (ECOWS). IEEE Computer<br />

Society, 2010, pp. 67–74.<br />

[5] D. Schuller, A. Polyvyanyy, L. García-Bañuelos, and<br />

S. Schulte, “Optimization of Complex QoS-Aware Service<br />

Compositions,” in Int. Conf. on Service Oriented Computing<br />

(ICSOC). Springer, 2011, pp. 452–466.<br />

[6] D. Schuller, A. Miede, J. Eckert, U. Lampe, A. Papageorgiou,<br />

and R. Steinmetz, “QoS-based Optimization of Service Compositions<br />

for Complex Workflows,” in Int. Conf. on Service<br />

Oriented Computing (ICSOC). Springer, 2010, pp. 641–648.<br />

[7] H. Taha, Operations Research – An Introduction, 8th ed.<br />

Pearson Prentice Hall, London, 2007.<br />

[8] U. Lampe, D. Schuller, J. Eckert, and R. Steinmetz, “Optimizing<br />

Service Selection for Probabilistic QoS Attributes,”<br />

in Workshop on Architecture, Concepts and Technologies for<br />

Service Oriented Computing (ACT4SOC). SciTePress, 2010,<br />

pp. 42–51.<br />

[9] C. W. Dawson and R. J. Dawson, “Generalised Activity-onthe-Node<br />

Networks for Managing Uncertainty in Projects,”<br />

Int. Journal of Project Management, vol. 13, no. 6, pp. 353–<br />

362, 1995.<br />

[10] R. J. Schonberger, “Why Projects are ’always’ late: a Rationale<br />

based on Manual Simulation of a PERT/CPM Network,”<br />

Interfaces, vol. 11, no. 5, pp. 66–70, 1981.<br />

[11] S. Rosario, A. Benveniste, S. Haar, and C. Jard, “Probabilistic<br />

QoS and Soft Contracts for Transaction-Based Web Services<br />

Orchestrations,” IEEE Transactions on Services Computing<br />

(TSC), vol. 1, no. 4, pp. 187–200, 2008.<br />

[12] A. Miede, U. Lampe, D. Schuller, J. Eckert, and R. Steinmetz,<br />

“Evaluating the QoS Impact of Web Service Anonymity,” in<br />

European Conf. on Web Services (ECOWS). IEEE Computer<br />

Society, 2010, pp. 75–82.<br />

[13] M. McLaughlin, “A Compendium of Common Probability<br />

Distributions,” http://www.causascientia.org/math stat/Dists/<br />

Compendium.html, 2001.<br />

[14] F. Lécué, “Optimizing QoS-Aware Semantic Web Service<br />

Composition,” in Int. Semantic Web Conf. (ISWC). Springer,<br />

2009, pp. 375–391.<br />

[15] N. B. Mabrouk, N. Georgantas, and V. Issarny, “A Semantic<br />

End-to-End QoS Model for Dynamic Service Oriented Environments,”<br />

in Workshop on Principles of Engineering Serviceoriented<br />

Systems (PESOS). IEEE Computer Society, 2009,<br />

pp. 34–41.<br />

[16] L. Zeng, B. Benatallah, A. H. H. Ngu, M. Dumas,<br />

J. Kalagnanam, and H. Chang, “QoS-Aware Middleware for<br />

Web Services Composition,” IEEE Transactions on Software<br />

Engineering (TSE), vol. 30, no. 5, pp. 311–327, 2004.<br />

[17] M. Li, J. Huai, and H. Guo, “An Adaptive Web Services<br />

Selection Method Based on the QoS Prediction Mechanism,”<br />

in Int. Joint Conf. on Web Intelligence and Intelligent Agent<br />

Technology (WI-IAT). ACM, 2009, pp. 395–402.<br />

[18] S.-Y. Hwang, H. Wang, J. Tang, and J. Srivastava, “A probabilistic<br />

approach to modeling and estimating the qos of webservices-based<br />

workflows,” Information Sciences (ISCI), vol.<br />

177, no. 23, pp. 5484–5503, 2007.<br />

[19] A. Bellucci, V. Cardellini, V. D. Valerio, and S. Iannucci, “A<br />

Scalable and Highly Available Brokering Service for SLA-<br />

Based Composite Services,” in Int. Conf. on Service Oriented<br />

Computing (ICSOC). Springer, 2010, pp. 527–541.<br />

[20] P. Leitner, W. Hummer, and S. Dustdar, “Cost-Based Optimization<br />

of Service Compositions,” IEEE Transactions on<br />

Services Computing (TSC), 2011, prePrint.


Research Article<br />

Mindfully resisting the bandwagon:<br />

reconceptualising IT innovation assimilation<br />

in highly turbulent environments<br />

Martin Wolf, Roman Beck, Immanuel Pahlke<br />

Institute of Information Systems, Goethe University Frankfurt, Frankfurt am Main, Germany<br />

Correspondence:<br />

M Wolf, Institute of Information Systems, Goethe University Frankfurt, Grüneburgplatz 1, Frankfurt am Main 60323,<br />

Germany.<br />

Tel: þ 49 (0)69 798 33862;<br />

Fax: þ 49 (0)69 798 33910;<br />

E-mail: mwolf@wiwi.uni-frankfurt.de<br />

Abstract<br />

Environmental turbulence (ET), as exemplified by the recent financial crisis between 2007<br />

and 2009, leads to a high degree of uncertainty, and fosters mimicry and resulting<br />

bandwagon phenomena in information technology (IT) innovation assimilation processes.<br />

In these highly turbulent environments, ‘mindless’ IT innovation assimilation by<br />

participating organizations plays a major role in the manifestation and facilitation of<br />

mimetic influences. Even in less turbulent economic cycles, highly turbulent industries<br />

such as the financial services industry have to deal with demanding IT innovation<br />

assimilation processes, and are exposed to varying levels of ET and mimicry. <strong>Dr</strong>awing<br />

upon the theory of dynamic capabilities, organizational mindfulness (OM) is one viable<br />

means to mitigate the potentially negative consequences of mimetic behaviour. Here,<br />

mindful organizations are more successful in overcoming situations of high dynamism, and<br />

sometimes are even able to exploit them. So far, little empirical research has been<br />

conducted to quantify the influence of OM in scenarios of high dynamism and mimicry. On<br />

the basis of 302 complete responses from senior IT managers in the financial services<br />

industry from the Anglo-Saxon countries (the United States, Canada and the United<br />

Kingdom), this research contributes to a deeper understanding of the interaction of OM<br />

with institutional pressures against the background of ET.<br />

Journal of Information Technology advance online publication, 5 June 2012;<br />

doi:10.1057/jit.2012.13<br />

Keywords: organizational mindfulness; IT innovation assimilation; institutional theory; bandwagon<br />

phenomena; IT business value; environmental turbulence<br />

Introduction<br />

Economic scenarios of high dynamism and volatility,<br />

such as the financial crisis between 2007 and 2009,<br />

demand a firm’s continuous technological adaptation<br />

to retain a competitive position and comply with regulatory<br />

influences (Barua et al., 2004; Sugumaran et al., 2008; Dong<br />

et al., 2009). In particular, scenarios of high uncertainty<br />

resulting from above-average volatility facilitate the emergence<br />

of mimicry and bandwagon phenomena among<br />

competing organizations that, in turn, might eventually<br />

negatively affect the realization of business value (Fiol<br />

and O’Connor, 2003; Swanson and Ramiller, 2004). For<br />

instance, uncertainty resulting from incomplete information<br />

on future market developments most likely leads to a<br />

Journal of Information Technology (2012), 1–23<br />

& 2012 JIT Palgrave Macmillan All rights reserved 0268-3962/12<br />

palgrave-journals.com/jit/<br />

situation where organizations inconsiderately follow a<br />

so-called ‘best practice approach’. By doing so, they may<br />

join an information technology (IT) innovation assimilation<br />

bandwagon generated by prior adopters. In such<br />

situations, organizations tend to justify their decisions with<br />

the consensus of the ‘herd’, rather than an environmentally<br />

aligned and customized-IT innovation strategy (Fiol and<br />

O’Connor, 2003). In this context, organizational mindfulness<br />

(OM) is assumed to be an effective means to identify<br />

and accommodate changes facilitated by the market, as well<br />

as to actively resist bandwagon phenomena that might<br />

otherwise negatively affect the generation of (IT-induced)<br />

business value (Fiol and O’Connor, 2003; Swanson and


2<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Ramiller, 2004; Butler and Gray, 2006). Bandwagons are<br />

defined as diffusion processes that are reflected by the<br />

individual or organizational adoption of an idea, technique,<br />

technology or product solely as a result of the number of<br />

organizations that have already adopted it (Abrahamson<br />

and Rosenkopf, 1990).<br />

As one specific instance of a complex and thus<br />

demanding IT innovation, the study at hand analyses the<br />

assimilation of Grid-based architectures (Foster and<br />

Kesselman, 1999; Buyya et al., 2009). In essence, Grid-based<br />

architectures serve to meet the volatile IT resource and<br />

IT service demands of organizations in highly turbulent<br />

environments (Hackenbroch and Henneberger, 2007).<br />

Following the seminal definition by Foster (2002), a Grid<br />

is a system that coordinates IT resources that are not<br />

subject to centralized control; uses standards, open<br />

protocols and interfaces; and delivers non-trivial qualities<br />

of service. Accordingly, Grid computing enables heterogeneous<br />

and geographically dispersed IT resources to be<br />

virtually shared and accessed across an industry, organization<br />

or workgroup. Increasingly, large-scale enterprise<br />

applications are no longer running on dedicated, centralized<br />

computing facilities. Instead, they operate on heterogeneous<br />

Grid resources that may span multiple<br />

administrative units across different locations within an<br />

organization (Strong, 2005). Depicting the continuous<br />

growth of the associated market for virtualized IT resource<br />

services, the market research company IDC (2010) projects<br />

that, with regard to revenues, the market will grow by<br />

27.4% per annum up to 2014. The report estimates that<br />

worldwide revenues from virtualized IT resource provisioning<br />

will grow from US$16 billion in 2009 to $55.5 billion<br />

in 2014 (IDC, 2010). Recently, the core concepts of Grid<br />

computing have transitioned to the domain of Cloud<br />

computing (Foster et al., 2008; Weinhardt et al., 2009),<br />

which represents a paradigmatic change in the dynamic<br />

provisioning of IT resources (Gartner, 2010).<br />

In general, IT innovations can be classified, according to<br />

Swanson (1994), into Type I innovations, which are purely<br />

technologically driven (e.g., database systems), and Type II<br />

innovations, which involve the technological support of<br />

administrative tasks (e.g., payroll or human resources<br />

systems). Since Grid-based architectures will eventually<br />

foster the development of advanced data processing, data<br />

mining and (inter-)organizational collaboration capabilities,<br />

they have the potential to generating strategic value<br />

through integration with firms’ core business model. They<br />

can therefore be classified as Type III innovations, which<br />

potentially can affect the entire business strategy, and thus<br />

are especially susceptible to bandwagon phenomena<br />

because of their strategic complexity. Despite their strategic<br />

potential, Type III innovation studies are still rarely<br />

covered by the existing literature. In particular, there has<br />

been little empirical research to explicate and quantify the<br />

interplay of mimetic pressure (MP) and resulting bandwagon<br />

phenomena influencing IT innovation assimilation that<br />

stem from environmental turbulence (ET) (Fiol and<br />

O’Connor, 2003; Pavlou and El Sawy, 2006). Furthermore,<br />

few studies have focused on dynamic capabilities, such as<br />

OM, which help to mitigate negative consequences of<br />

bandwagons. OM is defined by a firm’s ‘rich awareness of<br />

discriminatory detail and a capacity for action’ (Weick<br />

et al., 1999:37). OM is one means by which organizations<br />

can successfully overcome critical events, and master<br />

scenarios of high uncertainty, such as complex IT innovation<br />

processes. Such processes exhibit a high probability<br />

of negative consequences, such as significant monetary<br />

losses.<br />

In the context of IT innovation assimilation, OM is<br />

assumed to foster the identification of and resistance to<br />

purely mimetic IT assimilation behaviour (Fiol and<br />

O’Connor, 2003; Swanson and Ramiller, 2004). Moreover,<br />

as an outgrowth of a dynamic capability, OM might help<br />

organizations to cope better with ET and the resulting<br />

high uncertainty (Weick and Sutcliffe, 2007), eventually<br />

leading to above-average IT-based business value generation.<br />

Despite several calls to integrate the interplay between<br />

OM and MP in IT innovation assimilation processes<br />

(Fichman, 2004; Swanson and Ramiller, 2004), to the best<br />

of our knowledge this issue has not been addressed so<br />

far. Accordingly, our research conceptualizes this interplay<br />

as one instance of the institutional pressures driving the<br />

assimilation of IT innovation and OM against the background<br />

of a highly turbulent environment, such as the<br />

recent financial crisis. The research model draws on<br />

institutional theory to account for mimetic influences,<br />

and on the theory of the dynamic capabilities of the firm to<br />

integrate OM.<br />

The remainder of the paper is structured as follows.<br />

First, the research questions and their relevance are<br />

introduced in the context of the overall research model.<br />

Second, the theoretical background that informed the<br />

research model is depicted, and the related hypotheses<br />

are introduced. Then the research method used for the<br />

field study conducted among senior Anglo-Saxon IT<br />

decision makers and the results of partial least squares<br />

(PLS) analyses of 302 complete responses are depicted.<br />

Finally, the paper concludes by illustrating the contributions<br />

of this analysis, and highlighting some further<br />

research opportunities.<br />

Interplay between institutional pressure and mindfulness in<br />

highly turbulent environments<br />

Our study analysed the Anglo-Saxon financial services<br />

industry during the 2007–2009 financial crisis, as a recent<br />

example of a highly turbulent environment. During this<br />

extraordinary period the financial industry exhibited a high<br />

extent of market volatility and uncertainty about further<br />

market developments. This uncertainty resulted from rapid<br />

changes in the market (e.g., rapid developments and<br />

uncertainties in the corporate bond, mortgage and derivative<br />

markets), and from concurrent technological demand<br />

(resulting from new regulatory requirements, such as the<br />

Basel II accord), which are subsumed by the concept of<br />

ET (Pavlou and El Sawy, 2006). During this period, the<br />

highly turbulent market required organizations and financial<br />

services providers in particular, to assimilate IT<br />

innovations that were suitable to deal with these rapid<br />

changes. Concurrently, 172 US American banks failed,<br />

resulting in systemic bandwagon phenomena that exacerbated<br />

the extent of uncertainty and ET (FDIC, 2012). Even<br />

in times of low market volatility, the financial services<br />

industry is exposed to an above-average extent of


uncertainty, eventually leading to mimicry and bandwagon<br />

phenomena among competitors (Ang and Cummings,<br />

1997; Zhu et al., 2004). This uncertainty about future<br />

market developments and current market conditions might<br />

have serious negative effects on the generation and<br />

realization of business value from IT innovation assimilation.<br />

For instance, the adoption of an inadequate riskmanagement<br />

system might restrict the extent of regulatory<br />

compliance and thus impair the generation of IT-based<br />

business value. Consequently, our first guiding research<br />

question is:<br />

RQ1: How does environmental turbulence affect the<br />

influence of mimetic pressure and the realization of<br />

business value from IT innovation assimilation?<br />

Even in highly turbulent environments, some firms<br />

overcome related challenges more effectively than their<br />

competitors, or are potentially able to exploit them. This<br />

significant difference in IT business value generation can be<br />

attributed to advanced (dynamic) capabilities to align the<br />

IT innovation assimilation process to environmental contingencies.<br />

Organizations that exhibit enhanced dynamic<br />

capabilities (i.e., OM) can identify impending changes in<br />

the market earlier, and are able to derive highly contextualized<br />

IT innovation strategies. In addition, they can<br />

forecast and evaluate the consequences of the ensuing<br />

bandwagon phenomena. The second guiding research<br />

objective was therefore to assess the differences between<br />

rather mindful and less mindful firms in channelling MP to<br />

IT innovation assimilation processes against the background<br />

of ET. To achieve a holistic perspective on IT-based<br />

business value generation, we explicitly focused on the<br />

whole assimilation process, from initiation to routinization<br />

(Fichman, 2001), and thus omitted the non-assimilating<br />

firms, which had no expertise of Grid-based infrastructures.<br />

Consequently, we explore:<br />

RQ2: How do rather mindful and less mindful firms<br />

differ from each other with regard to the influence of<br />

mimetic pressure, and to the impact of environmental<br />

turbulence on IT-based business value generation stemming<br />

from its assimilation?<br />

Mimetic Pressure<br />

(MP)<br />

Coercive Pressure<br />

(CP)<br />

Normative Pressure<br />

(NP)<br />

Mindfully resisting the bandwagon M Wolf et al<br />

significant influence for both groups<br />

stronger influence for the more mindful<br />

than the less mindful group<br />

stronger influence for the less mindful<br />

than the more mindful group<br />

H1, H3<br />

Figure 1 IT innovation assimilation in highly turbulent environments.<br />

H2<br />

Top Management<br />

Support (TMS)<br />

To address RQ1 and RQ2, the paper proposes the<br />

research model depicted in Figure 1.<br />

The required dynamic capabilities encompass reasoning<br />

to actively resist prevailing bandwagon phenomena that<br />

stem solely from MP, rather than from rational decisionmaking.<br />

As far as the mediating agencies are concerned, we<br />

conceptualize the influence of institutional pressure on top<br />

management support (TMS) as the core human agency for<br />

channelling MP on the IT innovation assimilation process<br />

(Liang et al., 2007) against the moderating influence of ET.<br />

On the basis of 302 complete responses from organizations<br />

operating in a highly turbulent industry (i.e., the Anglo-<br />

Saxon financial services industry), our group comparisons<br />

of rather mindful and less mindful organizations indicate<br />

that OM indeed has a mitigating impact on the influence of<br />

MP and bandwagon phenomena in such an environment.<br />

Moreover, rather mindful organizations are able to benefit<br />

from ET, whereas less mindful firms are negatively affected<br />

with regard to business process performance (BPP).<br />

ET as driver of MP in IT innovation assimilation initiatives<br />

ET encompasses rapid environmental changes resulting<br />

from technological developments and changing market<br />

preferences (Pavlou and El Sawy, 2006; Buganza et al.,<br />

2009). Such changes eventually lead to uncertainty and<br />

unpredictability in market demand, consumer requirements<br />

and competitor strategies (Jap, 2001). Manifestations<br />

of ET (Buganza et al., 2009) eventually foster the emergence<br />

of uncertainty and mimicry among organizations, because<br />

of incomplete information about further market developments<br />

(Ang and Cummings, 1997; Fiol and O’Connor,<br />

2003). In order to mitigate such uncertainty, organizations<br />

are prone to imitate the behaviour of successful competitors<br />

in terms of their strategies for IT innovation<br />

assimilation (Swanson and Ramiller, 2004; Pavlou and El<br />

Sawy, 2006). Institutional theory provides one perspective<br />

to account for the rise of mimicry in highly turbulent<br />

environments (Meyer and Rowan, 1977; Ang and<br />

Cummings, 1997; Zhu et al., 2004; Liang et al., 2007).<br />

According to DiMaggio and Powell (1983), three different<br />

types of institutional pressure can be distinguished:<br />

mimetic, coercive and normative. In essence, a highly<br />

Environmental<br />

Turbulence (ET)<br />

H5<br />

Grid Assimilation<br />

(ASSM)<br />

H4<br />

Controls<br />

Grid Infrastructure Capabilities (GIC)<br />

Grid Technology Integration (GTI)<br />

Earliness of Grid Adoption (TIME)<br />

Firm Size (SIZE)<br />

Business Process<br />

Performance (BPP)<br />

3


4<br />

Mindfully resisting the bandwagon M Wolf et al<br />

turbulent environment, such as the financial services sector,<br />

makes an industry susceptible to mimetic behaviour (Fiol<br />

and O’Connor, 2003; Swanson and Ramiller, 2004). Thus,<br />

even if the consequences and benefits of an IT innovation<br />

are poorly understood, MP fosters its assimilation, if<br />

adopting firms are perceived as successful. By contrast,<br />

coercive pressure (CP) arises from societal expectations in<br />

a broader sense, where firms have to conform to expectations,<br />

policies or regulation from government, from<br />

customers or from the competitive environment. For<br />

example, one consequence of the financial crisis was tighter<br />

regulation of the financial services industry. Finally,<br />

normative pressure (NP) arises from the ongoing process<br />

of professionalization, which is further enforced by close<br />

collaboration with suppliers, business partners and governmental<br />

promotion. Since the focus of our study is on the<br />

interplay between mimicry and OM with regard to IT-based<br />

business value generation, we conceptualized normative<br />

and CPs as controls to account for other confounding<br />

institutional influences.<br />

In our case, mimicry is of special importance, particularly<br />

since organizations in the financial sector are exposed<br />

to an uncertain and hyper-competitive environment<br />

(Ang and Cummings, 1997). Consequently, structural<br />

and behavioural changes in firms are driven equally by<br />

an inherent organizational need for legitimacy and by<br />

considerations of competitive advantage and hidden<br />

efficiency potentials (Meyer and Rowan, 1977; DiMaggio<br />

and Powell, 1983). In some scenarios, information on the<br />

number and type of first adopters might outweigh the<br />

importance of the characteristics of the actual innovation<br />

(Meyer and Rowan, 1977; Aldrich and Fiol, 1994). Under<br />

the high-ambiguity conditions that prevail in highly<br />

turbulent environments, bandwagon pressure or MP is<br />

likely to be higher (Rosenkopf and Abrahamson, 1999).<br />

In the long run, this is likely to lead to organizational<br />

isomorphism and homogeneity (Meyer and Rowan, 1977;<br />

DiMaggio and Powell, 1983).<br />

Accordingly, we hypothesize:<br />

H1: Mimetic pressure drives top management to support<br />

IT innovation assimilation.<br />

H2: A higher extent of environmental turbulence<br />

strengthens the influence of mimetic pressure on top<br />

management support for IT innovation assimilation.<br />

OM as dynamic capability to accommodate conditions of high<br />

uncertainty<br />

The success of organizations that achieve and sustain a<br />

superior position in highly turbulent markets can be<br />

attributed, in particular, to their timely responsiveness<br />

(Teece et al., 1997). In this context, responsiveness can be<br />

characterized by the capability of continuous product<br />

innovation, complemented by management capability to<br />

coordinate and redeploy internal and external competences<br />

effectively with regard to identified changes in the<br />

environment (Teece et al., 1997). These characteristics are<br />

subsumed as part of the concept of dynamic capabilities<br />

(Teece et al., 1997), which captures a firm’s ability to tailor<br />

decisions and actions to environmental conditions, and<br />

identify mutual interdependencies (Lawrence and Lorsch,<br />

1967).<br />

In general, dynamic capabilities are defined as the<br />

‘capacity to renew competences so as to achieve congruence<br />

with the changing business environment’ and the ability to<br />

‘appropriately adapt, integrate, and reconfigure internal<br />

and external organizational skills, resources, and functional<br />

competences to match the requirements of a changing<br />

environment’ (Teece et al., 1997:515). These capabilities<br />

are critical in highly turbulent environments (Ang and<br />

Cummings, 1997). Here, accurate perception of the<br />

environment facilitates a better fit between decision making<br />

and the organizational context (Tenbrunsel et al., 1996).<br />

Moreover, dynamic capabilities foster the ability to identify<br />

new opportunities, to (re-)design business processes<br />

effectively and efficiently to foster business value generation<br />

(Teece et al., 1997), and to resist bandwagon<br />

phenomena that are inappropriate for the organizational<br />

context (Swanson and Ramiller, 2004). Highly turbulent<br />

environments, as exemplified by the financial services<br />

industry during the financial crisis, are characterized by a<br />

high extent of uncertainty, which demands a complex<br />

sense-making capability to safeguard organizational<br />

performance (McGill et al., 1993). Basically, the microfoundations<br />

of these sense-making capabilities can be<br />

attributed to scanning and interpretation activities (Thomas<br />

et al., 1993). Scanning is defined as ‘searching the external<br />

environment to identify important events or issues that<br />

might affect an organisation’ (Thomas et al., 1993:241). The<br />

scanning process is particularly critical for top management,<br />

which is usually exposed to more information than it can<br />

process in its decision making (Mintzberg, 1973). For<br />

instance, top management has to anticipate the consequences<br />

of a technological paradigm shift, such as the<br />

on-demand IT resource provisioning facilitated by Grid<br />

computing. In contrast, interpretation is defined as ‘development<br />

or application of ways of comprehending the<br />

meaning of information – it entails the fitting of information<br />

into some structure for understanding and action’ (Thomas<br />

et al., 1993:241). In the context of this research, interpretation<br />

involves the derivation of IT innovation strategies<br />

that maximize utilization of the benefits of Grid computing<br />

for the assimilating firm, and foster awareness of the<br />

inherent risks.<br />

In a first step, accurate and discriminant perception<br />

(scanning) and mindful evaluation (interpretation) of the<br />

environmental conditions explicitly allows for cognitive<br />

dissonance, which is assumed to be a core prerequisite for<br />

resisting bandwagon phenomena (Festinger, 1967; Fiol and<br />

O’Connor, 2003). Cognitive dissonance allows decision<br />

makers to notice contradicting information that might shed<br />

different light on future and past decisions (Festinger,<br />

1967). One concept that particularly explicates this<br />

cognitive state of heightened awareness and nuanced<br />

decision making is that of individual mindfulness (Langer,<br />

1989; Langer and Moldoveanu, 2000), which was eventually<br />

transferred by Weick et al. (1999) to the organizational<br />

level of high-reliability organizations (HROs). Analogous to<br />

financial services providers, these HROs, such as nuclear<br />

power plants and naval aircraft carriers, have to deal with<br />

unexpected events in an environment characterized by<br />

extreme turbulence, where error is omnipresent, and is


Mindfully resisting the bandwagon M Wolf et al<br />

most likely to have far-reaching consequences. Thus,<br />

these idiosyncratic requirements demand a sense-making<br />

capability that fosters continuous alignment with environmental<br />

conditions (Johns, 2006). Therefore, such organizational<br />

sense-making capabilities are not only important for<br />

the HRO domain, but also enable firms to stay efficient in<br />

highly turbulent environments, such as the financial<br />

market. During economic crises, a high extent of uncertainty<br />

and turbulence is channelled into the organizational<br />

IT innovation assimilation process (Fiol and<br />

O’Connor, 2003; Butler and Gray, 2006). Here, contextual<br />

information has to be gathered and aligned with the<br />

behaviour and decisions of competitors. In these scenarios,<br />

a mindful evaluation of contextual information improves a<br />

firm’s ability to resist or intentionally follow potential<br />

bandwagon phenomena to improve the business value<br />

gained from IT innovation assimilation (Fiol and O’Connor,<br />

2003). In sum, prior research has conceptualized mindfulness<br />

as both a cognitive capability of non-algorithmic<br />

thinking that can be consciously influenced by external<br />

stimuli, such as cognitive training sessions (Langer and<br />

Moldoveanu, 2000), and a static cognitive style that is stable<br />

and immutable (Sternberg, 2000). In the following, we<br />

assume that mindfulness is a cognitive capability that can<br />

be actively stimulated, and thus either evolves or deteriorates<br />

over time. According to the conceptualization of Weick<br />

and Sutcliffe (2007), OM is formed of five complementary<br />

cognitive dimensions that reflect the sub-dimensions of<br />

dynamic capabilities (i.e., scanning and interpretation):<br />

preoccupation with failure, reluctance to simplify, sensitivity<br />

to operations, commitment to resilience and deference to<br />

expertise.<br />

Preoccupation with failure defines a firm’s ability to learn<br />

from experiences in close-call situations, and to encourage<br />

proactive reporting and definition of mistakes (scanning).<br />

Consequently, it improves the firm’s learning capability,<br />

which is essential in crises (interpretation). The firm’s<br />

ability to ground its decision making on a more complete<br />

and nuanced picture of its operations, rather than draw<br />

from existing categories or solutions without considering<br />

contextual characteristics, is encompassed by its reluctance<br />

to simplify. Firms that exhibit a high extent of reluctance<br />

tosimplifyaremorelikelytoconsider(scanning) and<br />

integrate contradictory findings in their reasoning, and<br />

thus explicitly allow the interpretation of past practices as<br />

wrong (interpretation) (Fiol and O’Connor, 2003). In<br />

addition, such firms are more likely to attend to details<br />

based on organizational conditions (scanning), and exhibit<br />

the ability to extract the value of the information with<br />

respect to given circumstances (interpretation) (Fiol and<br />

O’Connor, 2003). The firm’s attentiveness and situational<br />

awareness to its operational front line are captured by its<br />

sensitivity to operations. With a high degree of sensitivity to<br />

operations, system anomalies, such as data anomalies,<br />

can be isolated while they are still tractable. This leads to a<br />

real-time processing of information (scanning) that can<br />

subsequently be considered for further decisions (interpretation).<br />

Commitment to resilience is defined by the<br />

firm’s ability to detect, contain and bounce back from<br />

inevitable errors to a dynamically stable state that facilitates<br />

the reconfiguration and transformation of business<br />

processes. Firms that have a well-developed commitment to<br />

resilience are more likely to learn through experimentation<br />

(scanning), and are thus able to interpret unusual and<br />

unexpected results in a more nuanced way (interpretation)<br />

(Fiol and O’Connor, 2003). Finally, the firm’s preference for<br />

systematically delegating decision-making processes to the<br />

most experienced employee, regardless of his or her hierarchical<br />

rank, defines the degree of deference to expertise.<br />

In essence, OM eventually leads to a state of high<br />

situational awareness and self-control. Accordingly, we<br />

propose that OM helps to actively resist bandwagon<br />

phenomena (i.e., MP) in situations where mimicry is not<br />

advantageous for the assimilating firm, and to make<br />

effective sense of a high degree of ET. Consequently, we<br />

hypothesize:<br />

H3: Rather mindful organizations are less affected by<br />

the impact of mimetic pressure on the IT innovation<br />

assimilation process compared with less mindful<br />

organizations.<br />

Consequences of mindful IT innovation assimilation at the<br />

business process level<br />

In contrast to the isolated decision to invest in an IT<br />

innovation (adoption), IT innovation assimilation is the<br />

continuous (and thus more holistic) organizational transition<br />

from a stage of technological initiation to a stage of<br />

routine utilization of the adopted IT innovation (Fichman,<br />

2001). Here, IT innovation assimilation is a process that<br />

can be characterized by seven stages (Fichman, 2001): in<br />

Stage 1 (initiation), an organization becomes aware of an IT<br />

innovation. The organization then decides to invest in this<br />

IT innovation (Stage 2: adoption), and to implement it<br />

(Stage 3: implementation). In successful IT innovation<br />

assimilation, the IT innovation is actively used in business<br />

processes (Stage 4: occasional usage to Stage 7: routinized<br />

usage).<br />

A business process level perspective was chosen as the<br />

unit of analysis for conceptualizing the progress of IT<br />

assimilation and business value generation, since IT<br />

investments are supposed to affect, first, the performance<br />

of specific business processes (Davamanirajan et al., 2006).<br />

Prior research suggests that a firm encompasses 18 key<br />

processes that are crucial for overall firm performance<br />

(Davenport, 1993). In order to identify the key business<br />

processes that are primarily influenced by Grid assimilation<br />

(ASSM) in the financial services industry, several expert<br />

interviews were conducted with IT executives. The interviews<br />

revealed that asset management, risk management<br />

and new product development are especially crucial<br />

processes for the financial services industry in this context.<br />

For instance, in 2008 Bank of America’s asset management<br />

process accounted for 10% of revenue but contributed<br />

35% to the net income (SEC, 2008), with growing projected<br />

potential for the future. Grid-based infrastructures foster<br />

assessment of the performance of asset allocation decisions<br />

in a more timely manner, thanks to concurrent processing<br />

(Hackenbroch and Henneberger, 2007). Further evidence of<br />

the relevance of the risk management process was provided<br />

by the Risk Management Association (RMA) in 2010.<br />

Among 75 global firms being interviewed by the RMA,<br />

56.8% rated the data quality of their risk management<br />

5


6<br />

Mindfully resisting the bandwagon M Wolf et al<br />

process as average or worse (RMA, 2010). The perception of<br />

the risk management professionals was that an improved<br />

data quality and risk management process would lead to a<br />

more timely identification of emerging problems, and more<br />

efficient capital allocation and utilization. Accordingly, the<br />

risk management process is essential, and vital in improving<br />

scanning capabilities, driven mainly by: (1) the pressure<br />

from regulators for better control of financial risks; (2) the<br />

globalization of financial markets, which has led to<br />

exposure to more sources of risk; and (3) technological<br />

advances that have made enterprise-wide risk management<br />

possible (Jorion, 2006). Grid-based infrastructures enable<br />

organizations to conduct more sophisticated risk management<br />

calculations, which lead to a gain of economic capital<br />

for follow-up businesses (Hackenbroch and Henneberger,<br />

2007). The need for continuous enhancement of the new<br />

product development process as a vital response capability<br />

is created mainly by fast-changing customer needs, which<br />

force financial services providers to provide highly customized<br />

financial products on demand. In such demanding<br />

environments, mindful organizations are able to realize<br />

more business value from IT innovation assimilation<br />

(Swanson and Ramiller, 2004). Mindful decision makers<br />

are able to efficiently identify growth options (to realize<br />

growth from an IT investment), deferral options (to<br />

postpone an IT investment until there is more information),<br />

learning options (to build on information gained<br />

from an initial investment) and staging options (to<br />

structure an IT innovation assimilation process into more<br />

manageable, separate process steps), which facilitate the<br />

realization of business value from IT innovation assimilation<br />

(Goswami et al., 2008). Here, OM manifests as<br />

contextually nuanced reasoning that facilitates the development<br />

of both an expanded scanning capability and an<br />

interpretation capability with regard to the utilized IT<br />

innovation (Goswami et al., 2008). Mindful organizations<br />

will therefore gather more relevant information, and are<br />

likely to be in a better position to recognize the various<br />

options that arise from IT innovation assimilation. In<br />

addition, OM eventually leads to an improved IT business<br />

alignment with regard to the fit of the IT innovation and<br />

overarching business objectives, and thus contributes to<br />

firm performance (Valorinta, 2009). Consequently, we<br />

hypothesize:<br />

H4: Rather mindful firms are able to realize a higher<br />

extent of business process performance resulting from IT<br />

innovation assimilation in the risk management, asset<br />

management and new product development process than<br />

less mindful firms.<br />

As an analytical perspective, real options pricing is<br />

especially appropriate for assessing IT innovation investment<br />

decisions under conditions of high uncertainty (Dixit<br />

and Pindyck, 1994), such as those that prevailed during the<br />

financial crisis. Mindfulness here facilitates differentiated<br />

reasoning with regard to the options that arise from IT<br />

innovation assimilation, based on the organization’s own<br />

facts and specifics (Goswami et al., 2008). In highly<br />

turbulent environments, the increased scanning and<br />

interpretation capabilities that characterize mindfulness<br />

lead to entrepreneurial alertness and the generation of<br />

digital options that, in turn, eventually enhance the<br />

generation of competitive actions and above-average firm<br />

performance (Sambamurthy et al., 2003). Hence, we<br />

hypothesize the following:<br />

H5: Rather mindful organizations are able to benefit<br />

from a high extent of environmental turbulence, whereas<br />

less mindful organizations are likely to be negatively<br />

affected by it.<br />

Controls<br />

IT innovation assimilation processes are subject to<br />

various other organizational influences (Fichman, 2001;<br />

Zhu et al., 2006b). To minimize the confounding impact of<br />

spurious correlation, we included Grid infrastructure<br />

capability (GIC) and Grid technology integration (GTI)<br />

adapted from Zhu et al. (2006b), firm size (SIZE), and<br />

earliness of Grid adoption (TIME) measured by years<br />

elapsed since first adoption (Fichman, 2001), as control<br />

variables for ASSM as our instance of an IT innovation and<br />

the realized business value through process performance<br />

improvements to account for differences among financial<br />

services providers.<br />

GIC captures the firm’s technical capability resulting<br />

from its having access to distributed computing power and<br />

purpose-specific technologies (e.g., a high-capacity, lowlatency<br />

network) within the firm. GTI refers to a set of<br />

investigation, evaluation and refinement activities aimed at<br />

creating a match between technological options and the<br />

application context (Iansiti, 1998). Consistent with the prior<br />

literature (e.g., Zhu and Kraemer, 2005; Zhu et al., 2006a),<br />

we assume that higher levels of GIC and GTI drive ASSM<br />

positively in the three identified key business processes.<br />

Moreover, higher levels of GIC and GTI constitute a vital<br />

prerequisite to benefiting from Grid infrastructure, resulting<br />

in performance improvements.<br />

Prior studies suggest that smaller firms are more flexible<br />

with regard to innovative technologies (Zhu and Kraemer,<br />

2005), but there is also an opposite theoretical view, based<br />

on the fact that larger firms are assumed to exhibit slack<br />

resources, which facilitate innovation diffusion (Tornatzky<br />

and Fleischer, 1990; Rogers, 1995). However, in general we<br />

assume that at least a certain size of firm is required for a<br />

Grid infrastructure to be implemented and utilized in a<br />

reasonable manner, since there has to be a significant<br />

number of IT resources (e.g., servers) that can be<br />

interconnected and virtualized. Consequently, we assessed<br />

responses only from financial services providers with more<br />

than 1000 employees. Among these, we hypothesize that the<br />

smaller firms are more likely to assimilate and profit from<br />

Grid technology, thanks to their higher openness to<br />

innovation. Finally, Grid experience reflects the fact that<br />

firms that initiated Grid implementation activities earlier<br />

had more time to reach later stages of IT innovation<br />

assimilation and create potential benefits (the so-called<br />

time lag effect).<br />

Empirical study<br />

In order to validate the research model, we conducted a<br />

questionnaire-based quantitative field study. The study was<br />

aimed at senior IT decision makers working for financial


Mindfully resisting the bandwagon M Wolf et al<br />

institutions located in the Anglo-Saxon countries (the<br />

United States, Canada and the United Kingdom) with more<br />

than 1000 employees. The research model was analysed<br />

using PLS, a component-based structural equation modelling<br />

(SEM) technique that concurrently assesses the<br />

psychometric properties of the measurement scales and<br />

the strength of the hypothesized relationships. We deemed<br />

this method appropriate for our research for several<br />

reasons. First, PLS handles measurement errors in exogenous<br />

variables better than other methods, such as multivariate<br />

regression (Chin, 1998). Second, component-based<br />

SEM approaches require fewer distributional assumptions<br />

with regard to the sample data (Cassel et al., 1999); in the<br />

early stages of measurement instrument development and<br />

theory testing, in particular, little is known about the<br />

distributional characteristics of the observed variables.<br />

Third, PLS can accommodate both exploratory and<br />

explanatory analyses, which is particularly suitable for<br />

our research context. Although PLS is often used for theory<br />

testing, it can provide first evidence for new relationships<br />

that are relevant for subsequent in-depth testing (Chin<br />

et al., 2003). Generally, the PLS approach is predictionoriented<br />

(Chin, 1998) and estimates latent variables as exact<br />

linear combinations of the observed measures (Wold,<br />

1982), which is advantageous, because theory generation<br />

is as important as theory testing.<br />

Data collection and sample profile<br />

A survey instrument was developed to collect the quantitative<br />

data required for model and hypothesis testing. On<br />

behalf of the authors, 2866 potential participants of a<br />

business panel were invited by a large international market<br />

research company to respond to the survey from August<br />

until September 2009. The date of invitation, the date of<br />

participation and the user ID were recorded to ensure that<br />

each panellist completed the online survey only once. In<br />

total, 782 responses were returned, indicating a response<br />

rate of 27.3%. Since the study was aimed at Gridassimilating<br />

organizations, the study participants were<br />

asked to indicate whether or not they had already<br />

assimilated Grid technology in at least one of the analysed<br />

processes. Responses from non-Grid assimilators, and<br />

responses that exhibited missing values, were removed.<br />

Thus, of the 782 completed responses, 349 were non-Grid<br />

assimilators and 131 questionnaires exhibited missing<br />

values and thus were excluded, which resulted in a sample<br />

of 302 complete responses.<br />

With respect to our research questions, the limited focus<br />

of our study on assimilating organizations is deemed<br />

appropriate, for the following reasons. First, the study at<br />

hand seeks to improve our understanding of the determinants<br />

of the assimilation stages of complex (i.e., Type III) IT<br />

innovations with respect to different levels of ET, interrelated<br />

institutional pressures and dynamic capabilities, such as<br />

OM. According to this research goal, we investigate the<br />

impact of variations in the assimilation stages of an IT<br />

innovation (e.g., Zhu and Kraemer, 2005; Liang et al., 2007)<br />

on BPP. We thereby follow the reasoning of Teo et al.<br />

(2003) that isomorphic processes influence all assimilation<br />

stages equally and eventually lead to further advance of<br />

organizations in the process. Second, assessing the impact of<br />

a Type III innovation on the process level (as our unit<br />

of analysis) requires an approach that focuses on how<br />

technology characteristics enhance specific business processes<br />

(Davamanirajan et al., 2006). Thus, only those firms<br />

that already had made sense of this technology (i.e., were at<br />

least already in the initiation stage), and had decided how<br />

to draw upon corresponding functionalities, were assumed<br />

to be able to realize benefits for the related business<br />

processes. In essence, we assume that Grid-based IT<br />

infrastructures have the potential to meet the volatile IT<br />

resource and service demands of organizations in highly<br />

turbulent environments. Consequently, we conducted a<br />

group comparison analysis, based on key performance<br />

indicators, between organizations that had and had not<br />

assimilated Grid-based infrastructures. In particular, we<br />

included an additional question in our survey asking for<br />

the name of the respondent’s firm. On the basis of this<br />

information, we were able to extract different key<br />

performance indicators for over 110 North American<br />

Organizations (including 39 non-assimilators and 71<br />

assimilators) from the BankScope database (30,000 companies<br />

listed) and the COMPUSTAT database (from<br />

Wharton Research Data Services; 25,000 companies listed).<br />

These databases encompass an extensive set of banks’<br />

financial statements, ratings and intelligence reports.<br />

For empirical investigation of the groups of assimilators<br />

and non-assimilators, we conducted several one-way<br />

analyses of variance (ANOVA), since this is a common<br />

approach to test for differences in indicators’ mean values<br />

(Box et al., 2005). In response to the need for statistical<br />

robustness, we also consider the results of Kruskal–Wallis<br />

rank-sum tests as the non-parametric version of ANOVA,<br />

and a generalized form of the Mann–Whitney approach<br />

(Kruskal and Wallis, 1952). The results depicted in<br />

Table A6 in the Appendix suggest that, at a 10% level of<br />

significance, firms that had assimilated Grid-based infrastructures<br />

for at least one of the identified key business<br />

process (i.e., asset management, risk management and new<br />

product development) were good as, or even better than,<br />

non-assimilating organizations with respect to asset quality,<br />

capital adequacy, profitability and efficiency and other<br />

liquidity indicators. Accordingly, we assume that nonassimilators<br />

have no business purpose for utilizing such<br />

an IT infrastructure, or they are not mindfully unaware of<br />

what the technology has to offer.<br />

Further details of the sample profile are depicted in<br />

Table 1.<br />

In the data-gathering stage, we followed a key informant<br />

approach (Bagozzi and Phillips, 1991) to collect data on<br />

Grid-related business processes of financial services providers.<br />

In IS research, this approach is often used to gather<br />

information on organizational factors, especially in the<br />

context of the business value of IT (Tanriverdi, 2005).<br />

However, if respondents and their context differ substantially<br />

from those who do not respond, then the respondents<br />

and the corresponding empirical findings may not be<br />

generalizable to the population (Miller and Smith, 1983).<br />

In order to address this issue of non-response bias, we<br />

compared the first 25% of responses with the last 25% (Sivo<br />

et al., 2006). Thereby, we follow Armstrong and Overton<br />

(1977), who argue that people responding in later waves<br />

can be assumed to be proxies for non-respondents. In this<br />

7


8<br />

Table 1 Sample profile<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Country Number of employees<br />

United States 189 (62.6%) 1001–5000 37 (12.3%)<br />

Canada 10 (3.3%) 5001–10,000 39 (12.9%)<br />

United Kingdom 103 (34.1%) 10,001–50,000 74 (24.5%)<br />

50,000+ 152 (50.3%)<br />

Respondent’s position Year of first Grid adoption<br />

CTO, COO, CIO 43 (14.2%) o2000 18 (6.0%)<br />

Chief systems architect 15 (5.0%) 2000–2001 20 (6.6%)<br />

Other senior IT decision maker 244 (80.8%) 2002–2003 16 (5.3%)<br />

2004–2005 38 (12.6%)<br />

2006–2007 93 (30.8%)<br />

2008–2009 117 (38.7%)<br />

regard, we treated the first 25% of the responses received<br />

as early respondents, and the last 25% of the responses as<br />

late respondents. We conducted both ANOVAs (parametric)<br />

and Kruskal–Wallis (non-parametric) tests to<br />

compare the responses of early and late respondents for<br />

all survey questions. We also compared firm differences<br />

with respect to key performance indicators extracted from<br />

archival data. The results depicted in Table A7 in the<br />

Appendix indicate that, at a 10% level of significance,<br />

almost no differences were found between the identified<br />

organizations in terms of asset quality, capital adequacy,<br />

profitability and efficiency or liquidity. In essence, the<br />

findings indicate that the responses of early and late<br />

respondents do not differ significantly, and thus indicate<br />

no substantial influence of non-response bias in our study.<br />

Research instrument: securing content validity<br />

Overall, our research model encompasses five reflective<br />

constructs (MP, CP, NP, TMS, ET) and two formative<br />

constructs (ASSM, BPP). All measures were informed by<br />

the extant literature, and were adapted to the Grid context<br />

where necessary (see Tables A1–A3 in the Appendix). To<br />

ensure content validity of the utilized measures, several<br />

expert interviews were conducted, and the survey instrument<br />

was made available to a panel of judges of both<br />

practitioners and academics (Straub, 1989; Straub et al.,<br />

2004).<br />

For the ASSM construct, a 7-item Guttmann scale was<br />

used to capture an organization’s current ASSM stage. This<br />

scale was based on prior research on the assimilation of<br />

software process innovations (Fichman, 2001), and on the<br />

assimilation of electronic procurement innovations (Rai<br />

et al., 2009). The respondents were requested to identify the<br />

current stage of ASSM for their risk management, asset<br />

management and new product development processes.<br />

As already outlined, these three processes take particular<br />

advantage of a high-performance Grid infrastructure,<br />

and were identified as being especially appropriate and<br />

vital for the financial services industry. Consequently, the<br />

measurement items of the assimilation construct focused<br />

on Grid-related activities in these processes.<br />

BPP was conceptualized as a dependent variable for each<br />

of the three key business processes, to capture the business<br />

value generation momentum of ASSM. As Karimi et al.<br />

(2007a) propose in the domain of enterprise resource<br />

planning assimilation, process efficiency, process effectiveness<br />

and process flexibility reflect the overall BPP<br />

construct. Process efficiency reflects the extent to which<br />

the use of IT innovation assimilation reduces operational<br />

costs and decreases the input/output conversion ratio, and<br />

process effectiveness is the extent to which IT innovation<br />

assimilation provides improved functionality, and enhances<br />

the quality of users’ work. The extent to which IT innovation<br />

assimilation provides firms with more flexibility<br />

in response to changing business environments defines<br />

the process flexibility dimension. Since successful business<br />

processes are characterized by the presence of<br />

efficiency, effectiveness and flexibility, which are not<br />

mutually interchangeable (Adler et al., 1999), we assume<br />

that the sub-dimensions of the BPP construct co-vary to a<br />

high extent (Karimi et al., 2007a, b). BPP was therefore<br />

operationalized as a second-order reflective model, and a<br />

two-stage approach was utilized (Yi and Davis, 2003) to<br />

model and include this construct as a dependent variable.<br />

The latent variable scores of the second-order construct<br />

were extracted in an initial analysis for each of the three<br />

identified business processes that were subsequently used<br />

as formative indicators for the estimation of the overall<br />

research model.<br />

In order to distinguish between firms with respect to<br />

their OM, we operationalized an OM score based on a<br />

second-order, Type II model. We therefore started off with<br />

wordings by Weick and Sutcliffe (2001) and Knight (2004),<br />

which we subsequently refined and adapted to create an<br />

initial pool of 25 items for the five distinct dimensions of<br />

OM (see Table A4 in the Appendix). To ensure content<br />

validity, we followed a two-staged approach. First, we<br />

conducted four expert interviews, which resulted in minor<br />

refinement, and the exclusion of two ambiguous items<br />

(Straub, 1989; Straub et al., 2004). Second, we conducted<br />

two rounds of unstructured and structured Q-sorting with<br />

different participants (four doctoral students and four<br />

experts from the financial services industry) (Moore and<br />

Benbasat, 1991). There was strong inter-judge reliability,<br />

and the required Cohen’s k of all constructs met the<br />

criterion of 0.65 (see Moore and Benbasat, 1991). Finally,<br />

we pre-tested the instrument before final rollout in a


small-scale setting with 37 respondents (2 chief operating<br />

officers, 5 chief technology officers, 4 chief information<br />

officers, 5 chief systems architects and 21 senior IT decision<br />

makers). Insufficient construct reliability meant that we<br />

had to drop 10 of the initial 25 (reflective) OM items.<br />

Because of the reflective second-order model specification<br />

(and thus assuming that items are interchangeable within<br />

their associated dimension), we deemed the impact of the<br />

dropped items on content validity as minor. Table A4 in<br />

the Appendix presents the final measurement scales of the<br />

different dimensions of OM.<br />

Analysis and results<br />

The results for the PLS estimation were obtained from<br />

SmartPLS (Version 2.0 M3; Ringle et al., 2005). The<br />

measurement and structural models were first evaluated<br />

separately for the rather mindful and less mindful groups,<br />

before the between-group comparisons were conducted.<br />

Consistent with Chin (1998), we utilized a 500 bootstrap<br />

sampling technique to test for the significance of the path<br />

estimates, factor loadings and weights.<br />

A variety of means for assessing between-group comparisons<br />

exist, including ANOVA and moderated regression<br />

(e.g., Sharma et al., 1981). In the realm of SEM, several<br />

means for testing between-group differences in these<br />

second-generation techniques have been developed (Qureshi<br />

and Compeau, 2009). In our particular case, we assess the<br />

between-group differences with PLS, as a component-based<br />

SEM approach, by following the widely used approach<br />

introduced by Chin (2000). This approach involves<br />

estimating model parameters for each group separately,<br />

using SmartPLS, and then performing a between-group test<br />

of significant differences (e.g., Keil et al., 2000; Venkatesh<br />

and Morris, 2000; Zhu et al., 2006a; Hsieh et al., 2008).<br />

We partition the sample on the basis of a median split of<br />

the calculated OM score (mean ¼ 5.29; median ¼ 5.41;<br />

Table 2 Descriptive statistics, and validity and reliability criteria<br />

Mindfully resisting the bandwagon M Wolf et al<br />

SD ¼ 1.03), resulting in a group of less mindful organizations<br />

(n ¼ 152) and a group of rather mindful organizations<br />

(n ¼ 150). In order to test for significant differences<br />

between the groups, we conducted two-sample t-tests on<br />

the estimated coefficients obtained from the bootstrap<br />

processes for each group (Chin, 2000; Sarstedt et al., 2011).<br />

According to Chin (2000), this procedure for comparing<br />

multiple groups with pairwise t-tests is subject to several<br />

assumptions about the data and the measurement<br />

model: (1) each nested model considered has to reach an<br />

acceptable goodness of fit; (2) there should be measurement<br />

invariance; and (3) the data should not deviate significantly<br />

from normality. In order to address the first two issues<br />

we examine the construct reliability and construct validity<br />

for each nested model separately. The results are given in<br />

Tables 2 and 3, and indicate that both nested models<br />

achieve an appropriate goodness of fit, and support the<br />

assumption of measurement invariance (i.e., no significant<br />

differences of the item loadings for each construct). To test<br />

for normal distribution of the estimations across all<br />

bootstrapping samples we performed tests of normality<br />

based on skewness and kurtosis (see Table A5 in the<br />

Appendix) and visually inspected q–q plots (Chambers<br />

et al., 1983; D’Agostino et al., 1990). Finally, we conducted<br />

a Levene test statistic (Levene, 1960) to assess whether the<br />

parameters’ standard deviations differed significant across<br />

the groups, and then applied the appropriate t-test statistics<br />

according to the definitions of Chin (2000) and Sarstedt<br />

et al. (2011).<br />

In response to the need for robustness in statistical<br />

analysis, and with respect to the non-normality of some<br />

estimations (see Table A5 in the Appendix), we also applied<br />

the non-parametric bootstrapping approaches proposed by<br />

Henseler et al. (2009), Sarstedt et al. (2011) and Henseler<br />

(2012) for multigroup analyses in PLS. These approaches<br />

estimate the probability of path differences based on the<br />

outcomes of the bootstrap processes for each group fitting<br />

OM high Mean SD AVE CR a MP CP NP TMS ET ASSM a BPP a<br />

MP 4.90 1.24 0.89 0.96 0.94 0.94<br />

CP 5.23 1.24 0.75 0.90 0.83 0.53 0.87<br />

NP 4.88 1.22 0.74 0.89 0.82 0.04 0.71 0.86<br />

TMS 5.65 1.08 0.80 0.95 0.94 0.19 0.36 0.40 0.89<br />

ET 5.87 0.97 0.56 0.91 0.88 0.24 0.39 0.39 0.23 0.75<br />

ASSM a 4.69 1.86 NA NA NA 0.31 0.33 0.43 0.41 0.22 NA<br />

BPP a 4.21 2.08 NA NA NA 0.21 0.28 0.34 0.23 0.28 0.50 0.94<br />

OM low Mean SD AVE CR a MP CP NP TMS ET ASSMa BPPa MP 4.44 1.02 0.83 0.94 0.90 0.91<br />

CP 4.55 1.13 0.66 0.85 0.76 0.49 0.81<br />

NP 4.16 1.01 0.65 0.84 0.72 0.15 0.32 0.81<br />

TMS 4.22 1.27 0.80 0.95 0.94 0.33 0.29 0.25 0.89<br />

ET 5.41 0.91 0.55 0.85 0.81 0.15 0.19 0.00 0.04 0.74<br />

ASSM a<br />

3.96 1.77 NA NA NA 0.15 0.13 0.21 0.33 0.04 NA<br />

BPPa 3.04 2.12 NA NA NA 0.04 0.13 0.05 0.31 0.09 0.44 0.93<br />

a Formative measure.<br />

Mean value (Mean), standard deviation (SD), average variance extracted (AVE), composite reliability (CR), Cronbach’s a, correlations<br />

among constructs (off-diagonal) and square roots of average variance extracted (diagonal).<br />

9


10<br />

Table 3 Item loadings and cross-loadings<br />

Mindfully resisting the bandwagon M Wolf et al<br />

OM high MP CP NP TMS ET ASSM a BPP a OM low MP CP NP TMS ET ASSM a BPP a<br />

MP1 0.93 0.53 0.41 0.18 0.16 0.24 0.10 MP1 0.87 0.38 0.15 0.25 0.18 0.18 0.12<br />

MP2 0.95 0.44 0.44 0.18 0.25 0.31 0.21 MP2 0.95 0.48 0.17 0.27 0.10 0.15 0.08<br />

MP3 0.95 0.50 0.49 0.19 0.26 0.32 0.15 MP3 0.91 0.48 0.11 0.26 0.14 0.07 0.07<br />

CP1 0.48 0.80 0.67 0.25 0.26 0.37 0.28 CP1 0.29 0.64 0.33 0.09 0.16 0.15 0.07<br />

CP2 0.44 0.90 0.59 0.38 0.32 0.21 0.18 CP2 0.36 0.93 0.30 0.31 0.22 0.17 0.13<br />

CP3 0.46 0.88 0.60 0.29 0.42 0.30 0.29 CP3 0.56 0.84 0.22 0.22 0.17 0.03 0.11<br />

NP1 0.39 0.56 0.86 0.29 0.32 0.34 0.26 NP1 0.15 0.29 0.81 0.22 0.08 0.18 0.08<br />

NP2 0.43 0.62 0.91 0.40 0.37 0.33 0.29 NP2 0.22 0.24 0.82 0.18 0.04 0.24 0.06<br />

NP3 0.39 0.66 0.80 0.31 0.31 0.46 0.34 NP3 0.01 0.24 0.77 0.20 0.05 0.09 0.01<br />

TMS1 0.09 0.28 0.29 0.87 0.20 0.35 0.24 TMS1 0.22 0.26 0.17 0.88 0.01 0.28 0.25<br />

TMS2 0.11 0.33 0.38 0.91 0.25 0.34 0.25 TMS2 0.25 0.23 0.18 0.92 0.08 0.36 0.24<br />

TMS3 0.23 0.39 0.38 0.91 0.27 0.36 0.18 TMS3 0.33 0.26 0.21 0.90 0.10 0.32 0.33<br />

TMS4 0.14 0.29 0.34 0.94 0.18 0.39 0.20 TMS4 0.25 0.27 0.27 0.93 0.04 0.30 0.32<br />

TMS5 0.28 0.33 0.35 0.84 0.12 0.41 0.17 TMS5 0.23 0.28 0.28 0.83 0.06 0.20 0.23<br />

ET1 0.06 0.21 0.22 0.11 0.71 0.01 0.15 ET1 0.05 0.19 0.04 0.04 0.80 0.13 0.07<br />

ET2 0.24 0.33 0.34 0.22 0.80 0.10 0.16 ET2 0.03 0.19 0.09 0.08 0.63 0.01 0.02<br />

ET3 0.23 0.27 0.36 0.19 0.79 0.19 0.19 ET3 0.20 0.27 0.14 0.09 0.60 0.07 0.03<br />

ET4 0.23 0.31 0.27 0.08 0.73 0.08 0.13 ET4 0.12 0.34 0.12 0.10 0.59 0.04 0.04<br />

ET5 0.16 0.31 0.27 0.18 0.82 0.24 0.27 ET5 0.20 0.14 0.04 0.04 0.64 0.01 0.09<br />

ET6 0.18 0.34 0.34 0.26 0.81 0.31 0.34 ET6 0.10 0.03 0.00 0.04 0.77 0.01 0.09<br />

ET7 0.11 0.28 0.21 0.06 0.67 0.06 0.21 ET7 0.21 0.14 0.03 0.01 0.64 0.01 0.08<br />

ASSM1a 0.22 0.24 0.32 0.38 0.13 0.80 0.58 ASSM1a 0.08 0.03 0.11 0.23 0.03 0.71 0.46<br />

ASSM2 a<br />

0.28 0.32 0.38 0.28 0.26 0.80 0.33 ASSM2 a<br />

0.11 0.12 0.18 0.28 0.02 0.86 0.35<br />

ASSM3a 0.26 0.26 0.37 0.35 0.17 0.86 0.31 ASSM3a 0.19 0.19 0.21 0.27 0.06 0.79 0.20<br />

BPP1a 0.33 0.33 0.41 0.34 0.25 0.68 0.89 BPP1a 0.08 0.15 0.27 0.11 0.60 0.19 0.90<br />

BPP2 a<br />

0.20 0.30 0.37 0.26 0.31 0.46 0.69 BPP2 a<br />

0.11 0.14 0.04 0.30 0.38 0.05 0.57<br />

BPP3a 0.22 0.26 0.38 0.24 0.34 0.54 0.77 BPP3a 0.15 0.14 0.12 0.22 0.55 0.18 0.87<br />

a Formative measure.<br />

Bold values refers to the associated construct, e.g., CP1 to CP3 refer to the CP construct.<br />

the PLS path modelling’s distribution-free characteristic.<br />

In particular, the approach proposed by Henseler et al.<br />

(2009) and Henseler (2012) reflects a Wilcoxon rank-sum<br />

test applied to the bootstrap values corrected for the<br />

original parameter values. In contrast, the approach<br />

proposed by Sarstedt et al. (2011) investigates the overlapping<br />

of bias-corrected bootstrap confidence intervals.<br />

In particular, we follow the so-called percentile method<br />

(Efron, 1981) for constructing appropriate bootstrapping<br />

confidence intervals accounting for the distribution’s<br />

skewness. We applied both non-parametric tests to the<br />

subsamples of our study, and derived P-values that are<br />

reported in Table 5 and are discussed in the corresponding<br />

section.<br />

Validation of the measurement model<br />

In order to validate the measurement model, the psychometric<br />

properties of all scales were assessed within the<br />

context of the structural model by examining the construct<br />

reliability and construct validity (Chin, 1998; Petter et al.,<br />

2007).<br />

Construct reliability refers to the internal consistency of<br />

the measurement model (Straub et al., 2004), or the degree<br />

to which items are free of systematic error, and yield<br />

consistent results. As our results in Table 2 indicate, the<br />

values of average variance extracted (AVE) of all constructs<br />

are above the recommended threshold of 0.5 (Fornell,<br />

1992). Consequently, more than 50% of measurement<br />

variance is explained by the constructs. Moreover, aggregate<br />

measures of the degree of inter-correlations among<br />

measurement items, such as the composite reliability or<br />

Cronbach’s a (a), exceed the recommended threshold of<br />

0.7 (Nunnally, 1978; Hair et al., 1998), suggesting internal<br />

consistency among the reflective measurement items. As far<br />

as the formative constructs (ASSM, BPP) are concerned,<br />

Table 4 shows that all weights are above 0.2, and almost<br />

all coefficients are statistically significant (Chin, 1998),<br />

indicating the relevance and consistency of the formative<br />

indicators used to measure these constructs. To ensure<br />

content validity, we decided to keep all indicators in our<br />

sample for the subsequent analyses (Petter et al., 2007).<br />

Furthermore, the estimated variance inflation factors (VIF)<br />

for the ASSM and BPP items (see Table 4) are all below 3.3<br />

(tolerance40.30), indicating no serious concern of multicollinearity<br />

(Petter et al., 2007).<br />

Construct validity captures whether the indicators of a<br />

construct measure what they are supposed to measure,<br />

from a psychometric perspective. Hence, the parameter<br />

estimates of the relationships between constructs and their<br />

indicators are assessed for consistency. Construct validity<br />

encompasses an assessment of both convergent validity and


Mindfully resisting the bandwagon M Wolf et al<br />

Table 4 Coefficients, t-values and VIF of exogenous variable TMS and formative measures ASSM and BPP<br />

OM high Coefficient t-value VIF Tolerance OM low Coefficient t-value VIF Tolerance<br />

MP 0.01 0.21 1.44 0.70 MP 0.23 2.64 1.40 0.71<br />

CP 0.14 1.31 2.36 0.42 CP 0.12 1.38 1.68 0.60<br />

NP 0.28 2.38 2.22 0.45 NP 0.14 1.82 1.17 0.85<br />

ET 0.08 1.05 1.27 0.78 ET 0.01 0.11 1.09 0.91<br />

ET MP 0.23 2.63 1.20 0.83 ET MP 0.32 4.83 1.10 0.91<br />

ASSM1 0.24 1.62 1.64 0.61 ASSM1 0.27 2.01 1.77 0.57<br />

ASSM2 0.28 1.93 1.56 0.64 ASSM2 0.50 2.66 1.73 0.58<br />

ASSM3 0.67 2.85 1.43 0.70 ASSM3 0.46 2.22 1.15 0.87<br />

BPP1 0.62 2.64 1.49 0.67 BPP1 0.53 2.37 1.73 0.58<br />

BPP2 0.33 1.84 1.48 0.68 BPP2 0.19 1.54 1.64 0.61<br />

BPP3 0.28 1.65 1.30 0.77 BPP3 0.48 2.29 1.24 0.81<br />

discriminant validity (Campbell and Fiske, 1959). The test<br />

for convergent validity is relevant only for reflective<br />

measures (MacKenzie et al., 2005), and determines whether<br />

the indicators of latent constructs that theoretically should<br />

be related are actually observed to be related. Loch et al.<br />

(2003) propose that the existence of significant interindicator<br />

and indicator-to-construct correlations suggests<br />

convergent validity; our results in Table 3 show that almost<br />

all loadings of the reflective constructs are greater than the<br />

recommended threshold of 0.707 (Chin, 1998), such that<br />

there exists more shared variance between the construct<br />

and its indicators than error variance, and the measurement<br />

items used are adequate for measuring the assigned<br />

constructs. Except for single items of the CP (low<br />

mindfulness group) and ET (both groups) constructs, all<br />

other items meet the recommended threshold. Consistently,<br />

the results in Table 3 reveal that the items load significantly<br />

higher on their own construct than on any other construct<br />

(cross-loadings). Furthermore, the corresponding loadings<br />

are higher than 0.5, suggesting the appropriateness of the<br />

items (Hulland, 1999). Consequently, we deemed this issue<br />

as minor for the measurement model quality.<br />

To assess the discriminant validity of the model specification,<br />

we tested whether indicators of latent constructs<br />

that are theoretically unrelated are indeed unrelated<br />

with regard to the parameter estimates. In this context,<br />

MacKenzie et al. (2005) propose an approach appropriate<br />

for evaluating the discriminant validity of formative and<br />

reflective measures, which analyses whether the interconstruct<br />

correlations are relatively low. The discriminant<br />

validity for the reflective constructs can be assessed by (1)<br />

analysing the cross-loadings and (2) assessing the Fornell–<br />

Larcker criterion. The cross-loadings indicate that each<br />

indicator exhibits a higher loading on its assigned construct<br />

than on the other constructs (Henseler et al., 2009), see<br />

Table 3. The Fornell–Larcker criterion (Fornell and Larcker,<br />

1981) postulates that a construct has to share more variance<br />

with its assigned indicators than with any other construct,<br />

as assessed by the relationships between the inter-construct<br />

correlations and the square roots of the AVE scores. From a<br />

psychometric perspective, the square root of the AVE for<br />

each construct should exceed the inter-construct correlations<br />

involving the construct (Fornell and Larcker, 1981).<br />

As depicted in Table 2, the square root of each AVE score is<br />

11<br />

greater than the correlations between the construct and any<br />

other construct, which indicates satisfactory discriminant<br />

validity.<br />

Validation of the structural model<br />

Since our analyses of the measurement model indicate<br />

convergent and discriminant validity, and since all<br />

indicators meet the established reliability and validity<br />

criteria, the presented measures were used to test the<br />

structural model and the associated hypotheses. The<br />

moderating effects of ET were operationalized and estimated<br />

following the procedure proposed by Chin et al.<br />

(2003). Accordingly, we first reduced multicollinearity by<br />

standardizing all indicators reflecting the predictor and<br />

moderator constructs to a mean of 0 and variance of 1. This<br />

step also allows for an easier interpretation of the path<br />

coefficient for the predictor variable. The path coefficient<br />

reflects the effect expected at the mean value of the<br />

moderator variable, which is 0. Second, using the standardized<br />

indicators of the predictor and moderator variables,<br />

we generated product indicators to reflect the latent<br />

interaction variables. Third, we utilized PLS to estimate<br />

the dependent variable TMS, with interaction terms as<br />

explanatory variables. Considering the potential high<br />

inter-correlations among the main effects and interaction<br />

terms, we also assessed the confounding influence of<br />

multicollinearity statistically according to the VIF. As we<br />

show in Table 4, the VIF scores of the explanatory variables<br />

of TMS are lower than the recommended level of 3.3<br />

(tolerance 4 0.30), which indicates the absence of multicollinearity<br />

(Petter et al., 2007).<br />

In Table 5, the validation results for the rather mindful<br />

firms and less mindful firms are depicted. We compared<br />

two nested models for the dependent variables ASSM and<br />

BPP to check for robustness of our results: Model 1 was the<br />

baseline model with the control factors only, and Model 2<br />

was the full model with all main effects, interaction effects<br />

and control effects. These models are fully nested, so that<br />

the difference in explanatory power allows a valid model<br />

comparison in terms of effect sizes. The explanatory power<br />

of the structural model is measured by the squared<br />

multiple correlations (R 2 ) of the dependent variables (Chin<br />

et al., 2003).


12<br />

Table 5 Empirical results<br />

Relationship OM low (n ¼ 152) OM high (n ¼ 150) Group comparison<br />

Basic model<br />

(controls only)<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Full<br />

model<br />

Basic model<br />

(controls only)<br />

As the results in Table 5 show, we find strong support<br />

for four of our five hypotheses (H1, H2, H3 and H5). Our<br />

analyses indicate that the moderating effect of ETs on the<br />

influence of MP on TMS for IT innovation assimilation<br />

initiatives and resulting IT business value differs for rather<br />

mindful and less mindful firms. If we consider the effect of<br />

MP on TMS for IT innovation assimilation initiatives, the<br />

results indicate that the coefficient of the relation of MP on<br />

TMS is not significant for the group of rather mindful firms<br />

(bMP-TMS ¼ 0.02, P40.05). However, the empirical results<br />

clearly suggest a significant positive effect of MP on TMS<br />

for the group of less mindful firms (b MP-TMS ¼ 0.23,<br />

Po0.01), which is positively moderated by ET (bET MP-TMS ¼<br />

0.32, Po0.01), indicating mimicry due to less-developed<br />

mindfulness in turbulent environments as a source of<br />

uncertainty. This effect decreases (bET MP-TMS ¼ 0.21,<br />

Po0.01) in the group of rather mindful organizations.<br />

Here, the interaction term (ET MP) indicates the strength<br />

of the theoretical relationship, and provides an estimate<br />

of the extent to which ET influences support for IT innovation<br />

assimilation. Moreover, by estimating the differences<br />

in the relevant IT innovation assimilation–performance<br />

Full<br />

model<br />

Coefficient<br />

(full model)<br />

Significance<br />

(parametric;<br />

Chin, 2000)<br />

Significance<br />

(non-parametric test)<br />

Henseler et al.<br />

(2009), Henseler<br />

(2012)<br />

Sarstedt<br />

et al.<br />

(2011)<br />

CP 0.23** 0.12* 0.17* 0.14 +0.02 0.86 0.45 40.10<br />

NP 0.17* 0.14* 0.27** 0.28** +0.14 0.18 0.10 o0.10<br />

MP — 0.23** — 0.02 0.21 0.04 0.03 o0.05<br />

ET — 0.01 — 0.08 +0.07 0.48 0.24 40.10<br />

ET MP — 0.32** — 0.21** 0.11 0.34 0.12 o0.10<br />

R2 (TMS) 0.11 0.24 0.17 0.23 — — — —<br />

DR2 — +0.13 — +0.06 — — — —<br />

GIC 0.10 0.05 0.03 0.04 0.09 0.96 0.47 40.10<br />

GTI 0.27** 0.32** 0.38** 0.30** 0.02 0.72 0.37 40.10<br />

TIME 0.09 0.08 0.13* 0.10 +0.02 0.83 0.41 40.10<br />

SIZE 0.22** 0.20** 0.03 0.03 +0.23 o0.01 o0.01 o0.025<br />

TMS — 0.31** — 0.32** +0.01 0.81 0.45 40.10<br />

R 2 (ASSM) 0.19 0.29 0.15 0.24 — — — —<br />

DR2 — +0.10 — +0.09 — — — —<br />

GIC 0.09 0.06 0.02 0.04 0.10 0.53 0.26 40.10<br />

GTI 0.17** 0.08 0.28** 0.02 0.10 0.68 0.32 40.10<br />

TIME 0.13 0.04 0.08 0.03 0.01 0.86 0.46 40.10<br />

SIZE 0.20** 0.06 0.05 0.02 +0.08 0.28 0.14 o0.05<br />

ASSM — 0.61** — 0.68** +0.07 0.56 0.29 40.10<br />

ET — 0.17** — 0.21** +0.38 o0.01 o0.01 o0.025<br />

R 2 (BPP) 0.13 0.46 0.10 0.56 — — — —<br />

DR2 — +0.33 — +0.46 — — — —<br />

P-values: *Po0.05; **Po0.01.<br />

The P-values, and the corresponding significance level for the group comparison where calculated, are based on bootstrap samples of<br />

1000 observations.<br />

Bold values refers to the associated construct, e.g., CP1 to CP3 refer to the CP construct.<br />

coefficients for the rather mindful and less mindful firms,<br />

the impact of this dynamic capability can be assessed. The<br />

results of the group comparison indicate that the strength<br />

of the direct link between ASSM and BPP increases in the<br />

group of mindful organizations (Db ASSM-BPP ¼þ0.07).<br />

Moreover, the results indicate that the effect of ET on<br />

BPP turns from a negative effect for the less mindful group<br />

(b ET-BPP ¼ 0.17, Po0.01) to a positive effect for the rather<br />

mindful firms (bET-BPP ¼ 0.21, Po0.01).<br />

With regard to the control variables, we note that SIZE<br />

seems to be negatively related to ASSM within the group of<br />

less mindful organizations (b SIZE-ASSM ¼ 0.20, Po0.01),<br />

whereas, in particular, GTI is positively related to the<br />

assimilation stage (ASSM) in both groups (bSIZE-ASSM ¼<br />

0.30/0.32, Po0.01). The other control variables are insignificant<br />

with respect to the effect on ASSM and BPP. In<br />

sum, the estimates clearly indicate the robustness of the<br />

hypothesized relations when we control for other influence<br />

factors (GIC, GTI, SIZE, TIME). Furthermore, assessment<br />

of the coefficients of determination (R 2 ) indicates that<br />

the hypothesized main effects and positive moderation of<br />

ET contribute substantially to the explanatory power of our


esearch model. The R 2 scores for the dependent variable in<br />

the full model are close to 0.30 for ASSM and 0.50 for the<br />

BPP, indicating that our full model explains a moderate<br />

amount of variance of the dependent variable (Chin, 1998).<br />

In particular, the explained variance increases by about<br />

10% for ASSM and 35% for BPP for our full model,<br />

compared with the baseline model containing only the<br />

control factors. This indicates the strength of the theoretical<br />

relationships found in our analysis, and thereby provides a<br />

solid estimation of the degree to which ET, as a source of<br />

uncertainties, encourages mimetic behaviour and influences<br />

both the adoption decision and assimilation process<br />

and the corresponding business value creation in our<br />

investigation.<br />

Since we obtained our survey data from key respondents<br />

in IT departments of financial services providers, and all selfreported<br />

data can potentially be confounded by common<br />

method bias (Podsakoff et al., 2003), we used several means<br />

to assess and minimize the presence of this potential threat.<br />

First, our measurement instrument contains different scales<br />

to reduce scale commonality (Podsakoff et al., 2003). After<br />

data collection, we conducted Harman’s one-factor test<br />

(Podsakoff and Organ, 1986) to control for single-respondent<br />

bias. An unrotated principal components factor analysis<br />

combining independent and dependent variables indicates<br />

that the identified principal factors account for only 22.3%<br />

of the covariance among all constructs, well below the 50%<br />

rule-of-thumb cut-off (Podsakoff and Organ, 1986). Next,<br />

following the recommendation of Podsakoff et al. (2003) and<br />

the analytical procedure used by Liang et al. (2007), we<br />

added a common method factor to the PLS measurement<br />

model. The indicators of all constructs were associated<br />

reflectively with this method factor. Then each indicator’s<br />

variances substantively explained by the principal construct<br />

and by the method factor were computed. The results<br />

show that (1) only one out of the 26 method loadings is<br />

significant, and (2) whereas the average substantively<br />

explained variance for an indicator is 0.716, the commonmethod-based<br />

variance is only 0.0026. In addition, results of<br />

the structural models demonstrated different levels of<br />

significance for path coefficients. On the basis of the results<br />

Table 6 Hypothesis testing results<br />

Mindfully resisting the bandwagon M Wolf et al<br />

of these tests, we concluded that there is no serious concern<br />

of common method bias for this study.<br />

Discussion<br />

Key findings<br />

In total, the results support four of the five hypotheses<br />

(see Table 6), and suggest that four relationships differ<br />

between the rather mindful group and the less mindful<br />

group (see Table 5). Consistent with Ang and Cummings<br />

(1997) and Liang et al. (2007), our results emphasize that,<br />

in particular, it is MP that drives top management to<br />

support IT innovation assimilation initiatives (H1) (see<br />

Figure 2). The behaviour of successful competitors is likely<br />

to initiate a new bandwagon that seduces other firms in the<br />

same market to join it without considering their firmspecific<br />

circumstances. Our results indicate that the<br />

influence of MP (as the primary origin of bandwagons)<br />

on TMS is indeed strengthened by a highly turbulent<br />

environment, such as the financial services industry (H2).<br />

Thereby, we underpin the conceptual work by Rosenkopf<br />

and Abrahamson (1999), who note that ET eventually<br />

fosters uncertainty and resulting mimicry. In sum, the<br />

underlying mechanisms seem to be similar in highly<br />

turbulent environments. In essence, our results reveal<br />

mechanisms of mimicry and herding behaviour that are<br />

potentially present in highly turbulent industries in general.<br />

With the distinction between rather mindful and less<br />

mindful firms we provide a more nuanced perspective<br />

on IT innovation assimilation and its exposure to ET, as<br />

well as the consequences of MP (H3, H4, H5). By doing so,<br />

we address calls by Swanson and Ramiller (2004), Butler<br />

and Gray (2006) as well as Fichman (2004) to assess the<br />

role of OM as a firm’s capability to resist misguiding<br />

environmental influences due to strict organization-specific<br />

reasoning, nuanced appreciation of the environment and<br />

contextualized derivation of the action repertoire. We find<br />

that OM mitigates the positive moderating impact of ET on<br />

the influence of TMS (H2). Consequently, in rather mindful<br />

firms top management is less likely to be affected by the MP<br />

Type of hypothesis Hypothesis Support<br />

Same impact across groups H1: Mimetic pressure drives the management to support IT innovation<br />

assimilation<br />

H2: A higher extent of environmental turbulence strengthens the<br />

influence of mimetic pressure on top management support for IT<br />

innovation assimilation<br />

Differential impact<br />

H3: Rather mindful organizations are less affected by the impact of<br />

across groups<br />

mimetic pressure on the IT innovation assimilation process compared<br />

with less mindful organizations<br />

H4: Rather mindful firms are able to realize a higher extent of business<br />

process performance resulting from IT innovation assimilation in the<br />

risk management, asset management and new product development<br />

process than less mindful firms<br />

H5: Rather mindful organizations are able to benefit from a high extent<br />

of environmental turbulence, whereas less mindful organizations are<br />

likely to be negatively affected by it<br />

Supported<br />

Supported<br />

Supported<br />

Not<br />

supported<br />

Supported<br />

13


14<br />

0.10<br />

**<br />

0.10<br />

**<br />

ns<br />

0.10<br />

Path coefficient for the more mindful group<br />

Path coefficient for the less mindful group<br />

Not significant<br />

0.10<br />

R values for the more mindful (left)<br />

and the less mindful (right) group<br />

Mimetic<br />

Coercive<br />

Normative<br />

H1, H3<br />

caused by ET. Moreover, the direct influence of MP on TMS<br />

may even vanish in rather mindful firms. This can be<br />

attributed to the capability of reflection in action, which is<br />

assumed to be especially developed in mindful organizations<br />

(Jordan et al., 2009). Reflection in action is defined by<br />

the ability to actively learn and realign from prior and<br />

current experiences, and in particular from critical,<br />

‘transformative’ change as initiated by bandwagons.<br />

Transformative change continuously challenges internalized<br />

ways of thinking, and therefore facilitates active<br />

resistance to bandwagons that are not appropriate for a<br />

firm-specific context.<br />

Interestingly, our results suggest that highly turbulent<br />

environments can be even an opportunity for rather<br />

mindful firms with regard to BPP (H4). In detail, our<br />

research indicates that rather mindful firms can benefit<br />

from ET at the business process level, whereas less mindful<br />

firms are negatively affected by rapid environmental<br />

changes with regard to BPP. This surprising result can be<br />

explained by the increased scanning and interpretation<br />

capabilities that rather mindful firms exhibit, which enable<br />

them to operate proactively in highly turbulent markets,<br />

rather than being passively driven by them (Teece, 2007).<br />

In contrast, less mindful firms are influenced by the<br />

institutional environment and prevailing turbulence, and<br />

therefore are more susceptible to bandwagon phenomena.<br />

This surprising result can be explained by the increased<br />

scanning and interpretation capabilities that rather mindful<br />

firms exhibit that enable them to operate proactively in<br />

highly turbulent markets (Teece, 2007). In the case of<br />

Grid computing, mindful firms are likely to identify the<br />

potential of this IT innovation to meet the varying IT<br />

resource demand in highly turbulent environments. In<br />

contrast, less mindful firms are driven by the institutional<br />

environment and prevailing turbulence and therefore are<br />

more susceptible to bandwagon phenomena. As far as the<br />

IT innovation assimilation process is concerned, less<br />

mindful firms are likely to assimilate it even if the IT<br />

innovation does not meet their specific requirements.<br />

ns<br />

0.23<br />

**<br />

ns<br />

0.12<br />

*<br />

0.28<br />

**<br />

0.23<br />

**<br />

Figure 2 Results of the group comparison (low vs high OM).<br />

P-values: *Po0.05; **Po0.01.<br />

Mindfully resisting the bandwagon M Wolf et al<br />

H2<br />

0.21<br />

**<br />

0.32<br />

**<br />

Environmental<br />

Turbulence<br />

Top Management<br />

Grid Assimilation<br />

H4<br />

Business Process<br />

Support 0.23 0.24<br />

0.32<br />

0.24 0.29<br />

0.68<br />

Performance 0.56 0.46<br />

**<br />

**<br />

0.31<br />

0.61<br />

**<br />

**<br />

H5<br />

0.21<br />

**<br />

-0.17<br />

**<br />

Controls<br />

Grid Infrastructure Capabilities (GIC)<br />

Grid Technology Integration (GTI)<br />

Earliness of Grid Adoption (TIME)<br />

Firm Size (SIZE)<br />

Finally, we found descriptive evidence that rather mindful<br />

firms can realize a higher level of business value from IT<br />

innovation assimilation at the business process level than<br />

can less mindful firms (H5). Nevertheless, with regard to<br />

inferential statistics, we could not find a significant<br />

difference. This can be explained by the fact that the<br />

realization of IT-based business value is highly contingent,<br />

and thus cannot be reduced solely to the IT innovation<br />

assimilation process itself. This would also explain why the<br />

direct effect of ET on business value generation is relatively<br />

strong and significant.<br />

Implications for research and practice<br />

The theoretical contribution of the depicted research is<br />

twofold. First, it contributes to the assimilation and<br />

diffusion of innovations theory by assessing the role of<br />

OM in mitigating the negative influences that potentially<br />

may arise from pure mimicry in the IT innovation<br />

assimilation process. By doing so, it sheds light on the<br />

micro-foundations of institutionalization and its interplay<br />

with the dynamic capabilities of the firm, which help to<br />

identify these influences and make the exposed firm aware<br />

of them. The firm is therefore enabled to consciously follow<br />

and frame a bandwagon or consciously resist it, rather than<br />

be passively driven by it. Our research thereby addresses<br />

recent calls to extend the nomological net of IT innovation<br />

assimilation as well as influencing and attenuating sources<br />

of institutionalization (e.g., Ang and Cummings, 1997; Fiol<br />

and O’Connor, 2003; Swanson and Ramiller, 2004). Second,<br />

our results indicate that OM is one viable means to realize a<br />

higher extent of IT business value against the background<br />

of a highly turbulent environment. Mindful firms are even<br />

potentially able to profit from ET. Consistent with the<br />

theory of dynamic capabilities, the discriminant perception<br />

of environmental change (scanning) and a response to it<br />

(interpretation) reflect an increased entrepreneurial alertness<br />

and the results of an above-average generation of<br />

digital options that are particularly valuable in a rapidly


Mindfully resisting the bandwagon M Wolf et al<br />

changing environment. Surprisingly, we find evidence that<br />

an environment of high turbulence can be an opportunity<br />

for rather mindful firms, whereas it reflects a threat for less<br />

mindful firms.<br />

In addition, our study on IT innovation assimilation<br />

(i.e., ASSM) has different implications for decision-makers<br />

of organizations with a high reliance on the availability and<br />

dependability of IT systems (such as financial services<br />

providers). First, senior IT decision-makers need to be<br />

aware of the level of participation and mindfulness that is<br />

required in the process of assimilating IT innovations<br />

(Liang et al., 2007). More importantly, they need to take<br />

into account that, with increased ET, they are exposed to<br />

higher MP. Because of uncertainty (e.g., incomplete<br />

information about future developments), firms tend to<br />

restructure themselves using other successful competitors<br />

in their market as role models. The awareness of this<br />

relation can help to improve scanning capabilities (e.g.,<br />

through better decision support systems) and initiate a<br />

mindful decision-making process to identify contextually<br />

appropriate IT innovation decisions. In this regard, OM<br />

as a dynamic capability and the corresponding complementary<br />

cognitive dimensions may serve as an overall<br />

organizational concept to ensure successful assimilation of<br />

IT innovation and business value generation in highly<br />

turbulent environments. OM can thereby be assumed to be<br />

one means to identify and accommodate changes facilitated<br />

by the market, and to resist arising bandwagon phenomena<br />

that might otherwise negatively affect the generation of ITinduced<br />

business value. In particular, mindfulness training<br />

sessions could empower employees to develop increased<br />

scanning and interpretation capabilities.<br />

Our research also informs the work of financial<br />

regulatory authorities, which should take a closer look at<br />

the mindful use of IT systems within the financial services<br />

industry. For example, one lesson learnt for regulators<br />

could be to enforce a stricter regime enforcing standards<br />

among less mindful organizations, or to be less supportive<br />

in crisis situation, in contrast with rather mindful<br />

competitors, in order to encourage the development of a<br />

rather mindful culture within the industry.<br />

Limitations and future research<br />

Despite the rich findings, our study has some limitations that<br />

suggest avenues for future research. In our study, we expand<br />

the nomological network of the assimilation and diffusion of<br />

innovation theory by assessing the role of OM as a dynamic<br />

capability mitigating the negative influences that potentially<br />

may arise from pure mimicry. We thereby expand the<br />

operationalization of cognitive organizational capabilities by<br />

aligning it to dynamic capabilities theory in the light of ETs.<br />

However, our study emphasizes the top management level as<br />

the sole human agency for transferring institutional pressures<br />

to the IT innovation assimilation process, although the extant<br />

literature suggests that the middle management level could<br />

be equally important for transferring such pressures (Liang<br />

et al., 2007). Accordingly, further research might find our<br />

research model a valuable starting point to integrate the<br />

middle management level as well.<br />

In addition, future research can investigate how<br />

other organizational capabilities influence the innovation<br />

15<br />

assimilation process in turbulent environments. The consequences<br />

of increasing ETs demand organizational sensemaking<br />

and responsiveness to safeguard organizational<br />

performance. Accordingly, companies can be seen as sensemaking<br />

units that are stimulated by ET, and which are<br />

constantly challenged to identify the appropriate contextual<br />

response (McGill et al., 1993). From a theoretical perspective,<br />

and with regard to the existing research in behavioural<br />

economics (e.g., Kahneman, 2003), future research should<br />

integrate and investigate factors to account for further<br />

cognitive capabilities and limitations, the amount and<br />

quality of accessible information, and other restrictions that<br />

influence an economic-rational decision-making process.<br />

In this regard, our theorizing is driven by an institutional<br />

perspective rather than by an adaptive learning perspective,<br />

although human beings (and therefore top managers) are<br />

constantly engaged in active sensing, responding and<br />

learning processes within different environmental conditions<br />

(Overby et al., 2006). Specifically, we focus on the<br />

relatively understudied relationship between uncertainties<br />

resulting from ETs and mimetic isomorphism that constitutes<br />

one of the core arguments in institutional theory<br />

(DiMaggio and Powell, 1983). Since our current sample<br />

encompasses organizational data, further research could<br />

explore the underlying processes at the individual level.<br />

Accordingly, future studies may employ a process view<br />

of isomorphism, with a special focus on the early stage<br />

of the innovation assimilation process, by investigating<br />

individual-level data from organizational decision makers<br />

facing critical make-or-buy decisions. Such longitudinal<br />

investigations could address recent calls by researchers<br />

for increased attention to the mechanisms and microfoundations<br />

of institutional theory (Powell and Colyvas,<br />

2008; Battilana et al., 2009), which have received only<br />

limited attention in the existing literature.<br />

From a methodological perspective, our sample consists<br />

of firms with more than 1000 employees, and thus<br />

potentially limits the generalizability of the results. Moreover,<br />

the responses of our survey represent the IT delivery<br />

side, which might overemphasize a technologically driven<br />

perspective. Dyadic data matching responses from the IT<br />

department with responses from the business department<br />

could overcome this limitation, and provide further<br />

insights into the process of assimilating IT innovation.<br />

Finally, our analyses focus on the financial services<br />

industry, which might be idiosyncratic with respect to an<br />

above-average level of ET. Concurrently, our findings<br />

represent an avenue for further research that could draw<br />

a more nuanced picture of the role of ETs and necessary<br />

dynamic capabilities in the innovation assimilation process.<br />

Conclusion<br />

In essence, ‘mindless’ IT innovation assimilation in highly<br />

dynamic environments, such as the financial services<br />

industry, can eventually lead to the emergence of bandwagons<br />

and a loss in BPP. A well-developed organizational<br />

scanning and interpretation ability is required to comprehend<br />

the implications of complex IT innovations and the<br />

relevant IT innovation decisions that implement them.<br />

In contrast, not all IT departments are able to ‘mindfully’


16<br />

derive the strategic implications of IT innovations, such as<br />

Grid computing.<br />

Our results indicate that OM is one way to contrast<br />

IT innovation strategies that eventually serve as a ‘fire<br />

accelerant’ with a mindful mitigation of bandwagon<br />

influences and the accompanying precise risk assessment.<br />

This mindfulness in the assimilation of IT innovation might<br />

not be restricted to the originating firm; it might also alert<br />

other, less mindful firms, since the overall market volatility<br />

is likely to be mitigated. Mindfulness analyses at an interorganizational<br />

or industry-level perspective could be a<br />

fruitful area for future research. In the end, OM could be<br />

propagated by the same means as the financial crisis itself:<br />

MP. A majority of mindful firms can force their less<br />

mindful competitors to follow.<br />

References<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Abrahamson, E. and Rosenkopf, L. (1990). When Do Bandwagon Diffusions<br />

Roll? How Far Do They Go? And When Do They Roll Backwards? A<br />

Computer Simulation, Academy of Management Best Paper Proceedings<br />

50(1): 155–159.<br />

Adler, P.S., Glodoftas, B. and Levine, D.L. (1999). Flexibility Versus Efficiency?<br />

A Case Study of Model Changeovers in the Toyota Production System,<br />

Organization Science 10(1): 43–68.<br />

Aldrich, H.E. and Fiol, C.M. (1994). Fools Rush In? The Institutional Context of<br />

Industry Creation, Academy of Management Review 19(1): 645–670.<br />

Ang, S. and Cummings, L.L. (1997). Strategic Response to Institutional<br />

Influences on Information Systems Outsourcing, Organization Science 8(3):<br />

235–256.<br />

Armstrong, J. and Overton, T. (1977). Estimating Non-Response Bias in Mail<br />

Surveys, Journal of Marketing Research 14(8): 396–402.<br />

Bagozzi, R.P. and Phillips, L.W. (1991). Assessing Construct Validity in<br />

Organizational Research, Administrative Science Quarterly 36(3): 421–458.<br />

Barua, A., Konana, P., Whinston, A.B. and Fang, Y. (2004). An Empirical<br />

Investigation of Net-Enabled Business Value, MIS Quarterly 28(4):<br />

585–620.<br />

Battilana, J., Leca, B. and Boxenbaum, E. (2009). How Actors Change<br />

Institutions: Towards a theory of institutional entrepreneurship, Academy of<br />

Management Annals 3(1): 65–107.<br />

Box, G.E.P., Hunter, J.S. and Hunter, W.G. (2005). Statistics for Experimenters:<br />

Design, innovation, and discovery, 2nd edn, New York: Wiley-Interscience.<br />

Buganza, T., Dell’Era, C. and Verganti, R. (2009). Exploring the Relationships<br />

Between Product Development and Environmental Turbulence: The case of<br />

mobile TLC services, Journal of Product Innovation Management 26(3):<br />

308–321.<br />

Butler, B.S. and Gray, P.H. (2006). Reliability, Mindfulness, and Information<br />

Systems, MIS Quarterly 30(2): 211–224.<br />

Buyya, R., Yeo, C.S., Venugopal, S., Broberg, J. and Brandic, I. (2009). Cloud<br />

Computing and Emerging IT Platforms: Vision, hype, and reality for<br />

delivering computing as the 5th utility, Future Generation Computer Systems<br />

25(6): 599–616.<br />

Campbell, D.T. and Fiske, D.W. (1959). Convergent and Discriminant<br />

Validation by the Multitrait-Multimethod Matrix, Psychological Bulletin<br />

56(1): 81–105.<br />

Cassel, C., Hackl, P. and Westlund, A. (1999). Robustness of Partial<br />

Least-Squares Method for Estimating Latent Variable Quality Structures,<br />

Journal<br />

of Applied Statistics 26(4): 435–446.<br />

Chambers, J.M., Cleveland, W.S., Kleiner, B. and Tukey, P.A. (1983). Graphical<br />

Methods for Data Analysis, Belmont, CA: Wadsworth.<br />

Chin, W.W. (1998). The Partial Least Squares Approach to Structural Equation<br />

Modeling, in G.A. Marcoulides (ed.) Modern Methods for Business Research,<br />

Hillsdale, NY: Lawrence Erlbaum Associates, pp. 295–336.<br />

Chin, W.W. (2000). Frequently asked questions: Partial least squares and<br />

PLS-graphs, [WWW document] http://disc-nt.cba.uh.edu/chin/plsfaq/<br />

plsfaq.htm (accessed 2 September 2011).<br />

Chin, W.W., Marcolin, B.L. and Newsted, P.R. (2003). A Partial Least Squares<br />

Latent Variable Modeling Approach for Measuring Interaction Effects:<br />

Results from a Monte Carlo simulation study and an electronic-mail<br />

emotion/adoption study, Information Systems Research 14(2): 189–217.<br />

D’Agostino, R.B., Belanger, A. and D’Agostino Jr, R.B. (1990). A Suggestion for<br />

Using Powerful and Informative Tests of Normality, American Statistician<br />

44(1): 316–332.<br />

Davamanirajan, P., Kauffman, R.J., Kriebel, C.H. and Mukhopadhyay, T.<br />

(2006). Systems Design, Process Performance, and Economic Outcomes in<br />

International Banking, Journal of Management Information Systems 23(2):<br />

65–90.<br />

Davenport, T. (1993). Process Innovation: Reengineering work through IT,<br />

Boston, MA: Harvard Business School.<br />

DiMaggio, P.W. and Powell, W.W. (1983). The Iron Cage Revisited:<br />

Institutional isomorphism and collective rationality in organizational fields,<br />

American Sociological Review 48(2): 147–160.<br />

Dixit, A.K. and Pindyck, R.S. (1994). Investment under Uncertainty, Princeton,<br />

NJ: Princeton University Press.<br />

Dong, S., Xu, S.X. and Zhu, K.X. (2009). Information Technology in Supply<br />

Chains: The value of IT-enabled resources under competition, Information<br />

Systems Research 20(1): 18–32.<br />

Efron, B. (1981). Nonparametric standard errors and confidence intervals,<br />

Canadian Journal of Statistics 9(2): 139–172.<br />

FDIC (2012). Failed Bank List, [WWW document] http://www.fdic.gov/bank/<br />

individual/failed/banklist.html (accessed: 18 April 2012).<br />

Festinger, L. (1967). A Theory of Cognitive Dissonance, Evanston, IL: Row<br />

Peterson.<br />

Fichman, R.G. (2001). The Role of Aggregation in the Measurement of<br />

IT-related Organizational Innovation, MIS Quarterly 25(4): 427–455.<br />

Fichman, R.G. (2004). Going Beyond the Dominant Paradigm for Information<br />

Technology Innovation Research: Emerging concepts and methods, Journal<br />

of the Association for Information Systems 5(8): 314–355.<br />

Fiol, C.M. and O’Connor, E.J. (2003). Waking Up! Mindfulness in the<br />

Face of Bandwagons, Academy of Management Review 28(1): 54–70.<br />

Fornell, C. (1992). A National Customer Satisfaction Barometer: The Swedish<br />

experience, Journal of Marketing 56(1): 6–21.<br />

Fornell, C. and Larcker, D.F. (1981). Evaluating Structural Equation Models<br />

with Unobservable Variables and Measurement Error: Algebra and statistics,<br />

Journal of Marketing Research 18(1): 382–388.<br />

Foster, I. (2002). What is the Grid? A Three Point Checklist, GRID Today 1(6):<br />

22–25.<br />

Foster, I. and Kesselman, C. (1999). The Grid: Blueprint for a new computing<br />

infrastructure, San Francisco, CA: Morgan Kaufmann.<br />

Foster, I., Zhao, Y., Raicu, I. and Lu, S. (2008). Cloud Computing and<br />

Grid Computing 360-Degree Compared, in Proceedings of the IEEE Grid<br />

Computing Environments Workshop (Austin, TX, USA, 2008); San Diego,<br />

CA: IEEE Computer Society Press.<br />

Gartner (2010). Forecast: Public Cloud Services, Worldwide and Regions,<br />

Industry Sectors, 2009–2014, [WWW document] http://www.gartner<br />

.com/DisplayDocument?ref ¼ clientFriendlyUrl&id ¼ 1378513 (accessed<br />

29 March 2011).<br />

Goswami, S., Teo, H.H. and Chan, H.C. (2008). Real Options from RFID<br />

Adoption: The role of institutions and managerial mindfulness, in<br />

Proceedings of the 29th Annual International Conference on Information<br />

Systems (ICIS) (Paris, France, 2008); Atlanta, GA: Association for<br />

Information Systems.<br />

Hackenbroch, W. and Henneberger, M. (2007). Grid Computing Infrastructures<br />

and their Value for Risk Management, in Proceedings of the 40th Annual<br />

Hawaii International Conference on System Sciences (HICSS) (Waikoloa,<br />

Hawaii, 2007); San Diego, CA: IEEE Computer Society Press.<br />

Hair, J., Anderson, R., Tatham, R. and Black, W. (1998). Multivariate Data<br />

Analysis with Readings, 4th edn, Englewood Cliffs, NJ: Prentice Hall.<br />

Henseler, J. (2012). PLS-MGA: A non-parametric approach to partial least<br />

squares-based multi-group analysis, in W.A. Gaul, A. Geyer-Schulz,<br />

L. Schmidt-Thieme and J. Kunze (eds.) Challenges at the Interface of<br />

Data Analysis, Computer Science, and Optimization, Berlin, Heidelberg:<br />

Springer, pp. 495–501.<br />

Henseler, J., Ringle, C.M. and Sinkovics, R.R. (2009). The Use of Partial<br />

Least Squares Path Modeling in International Marketing, in R.R. Sinkovics<br />

and P.B. Ghauri (eds.) Advances in International Marketing, Vol. 20, Bingley:<br />

Emerald JAI Press, pp. 277–319.


Mindfully resisting the bandwagon M Wolf et al<br />

Hsieh, J.J.P.-A., Rai, A. and Keil, M. (2008). Understanding Digital Inequality:<br />

Comparing continued use behavioral models of the socio-economically<br />

advantaged and disadvantaged, MIS Quarterly 32(1): 97–126.<br />

Hulland, J. (1999). Use of Partial Least Squares (PLS) in Strategic Management<br />

Research: A review of four recent studies, Strategic Management Journal<br />

20(2): 195–204.<br />

Iansiti, M. (1998). Technology Integration: Making critical choices in a<br />

dynamic world, Boston, MA: Harvard Business Press.<br />

IDC (2010). Through 2014 Public IT Cloud Services Will Grow at More Than<br />

Five Times the Rate of Traditional IT Products, New IDC Research Finds<br />

[WWW document] http://www.idc. com/about/viewpressrelease. jsp?<br />

containerId ¼ prUS22393210&sectionId ¼ null&elementId ¼ null&<br />

pageType ¼ SYNOPSIS (accessed 18 April 2012).<br />

Jap, S.D. (2001). Perspectives on Joint Competitive Advantages in Buyer–<br />

Supplier Relationships, International Journal of Research in Marketing<br />

18(1–2): 19–35.<br />

Jaworski, B.J. and Kohli, A.K. (1993). Market Orientation: Antecedents and<br />

consequences, Journal of Marketing 57(7): 53–70.<br />

Johns, G. (2006). The Essential Impact of Context on Organizational Behavior,<br />

Academy of Management Review 31(2): 386–408.<br />

Jordan, S., Messner, M. and Becker, A. (2009). Reflection and Mindfulness in<br />

Organizations: Rationales and possibilities for integration, Management<br />

Learning 40(4): 465–473.<br />

Jorion, P. (2006). Value at Risk, 3rd edn, Burr Ridge, IL: Richard D. Irwin.<br />

Kahneman, D. (2003). Maps of Bounded Rationality: Psychology for behavioral<br />

economics, American Economic Review 93(5): 1449–1475.<br />

Karimi, J., Somers, T.M. and Bhattacherjee, A. (2007a). The Impact of ERP<br />

Implementation on Business Process Outcomes: A factor-based study,<br />

Journal of Management Information Systems 24(1): 101–134.<br />

Karimi, J., Somers, T.M. and Bhattacherjee, A. (2007b). The Role of<br />

Information Systems Resources in ERP Capability Building and Business<br />

Process Outcomes, Journal of Management Information Systems 24(2):<br />

221–260.<br />

Keil, M., Tuunainen, V. and Wassenaar, A. (2000). A Cross-Cultural Study on<br />

Escalation of Commitment Behaviors in Software Projects, MIS Quarterly<br />

24(2): 299–325.<br />

Knight, A.P. (2004). Measuring Collective Mindfulness and Exploring its<br />

Nomological Network, Unpublished Master thesis, University of Maryland,<br />

College Park.<br />

Kruskal, W.H. and Wallis, W.A. (1952). Use of Ranks in One-Criterion<br />

Variance Analysis, Journal of the American Statistical Association 47(260):<br />

583–621.<br />

Langer, E.J. (1989). Minding Matters: The consequences of mindlessness–<br />

mindfulness, San Diego, CA: Academic Press.<br />

Langer, E.J. and Moldoveanu, M. (2000). The Construct of Mindfulness, Journal<br />

of Social Issues 56(1): 1–9.<br />

Lawrence, P.R. and Lorsch, J.W. (1967). Organization and Environment, Irwin,<br />

IL: Homewood.<br />

Levene, H. (1960). Robust Tests for Equality of Variances, in I. Olkin (ed.)<br />

Contributions to Probability and Statistics, Palo Alto, CA: Stanford University<br />

Press, pp. 278–292.<br />

Liang, H., Nilesh, S., Hu, Q. and Xue, Y. (2007). Assimilation of Enterprise<br />

Systems: The effect of institutional pressures and the mediating role of top<br />

management, MIS Quarterly 31(1): 59–87.<br />

Loch, K.D., Straub, D.W. and Kamel, S. (2003). Diffusing the Internet in the<br />

Arab World: The role of social norms and technological culturation, IEEE<br />

Transactions on Engineering Management 50(1): 45–63.<br />

MacKenzie, S.B., Podsakoff, P.M. and Jarvis, C.B. (2005). The Problem of<br />

Measurement Model Misspecification in Behavioral and Organizational<br />

Research and Some Recommended Solutions, Journal of Applied Psychology<br />

90(4): 710–730.<br />

McGill, A.R., Johnson, M.D. and Bantel, K.A. (1993). Cognitive Complexity<br />

and Conformity: The effects on performance in a turbulent environment,<br />

Academy of Management Proceedings 75(3): 379–383.<br />

Meyer, J. and Rowan, B. (1977). Institutional Organizations: Formal<br />

structure as myth and ceremony, American Journal of Sociology 83(2):<br />

340–363.<br />

Miller, L. and Smith, K. (1983). Handling Non-Response Issues, Journal of<br />

Extension 21(9–10): 45–50.<br />

Mintzberg, H. (1973). The Nature of Managerial Work, New York:<br />

Harper & Row.<br />

17<br />

Moore, G.C. and Benbasat, I. (1991). Development of an Instrument to Measure<br />

the Perceptions of Adopting an Information Technology Innovation,<br />

Information Systems Research 2(3): 192–222.<br />

Nunnally, J.C. (1978). Psychometric Theory, New York: McGraw-Hill.<br />

Overby, E., Bharadwaj, A.S. and Sambamurthy, V. (2006). Enterprise Agility<br />

and the Enabling Role of Information Technology, European Journal of<br />

Information Systems 15(2): 120–131.<br />

Pavlou, P.A. and El Sawy, O.A. (2006). From IT Leveraging Competence to<br />

Competitive Advantage in Turbulent Environments: The case of new product<br />

development, Information Systems Research 17(3): 198–227.<br />

Petter, S., Straub, D. and Rai, A. (2007). Specifying Formative Constructs in<br />

Information Systems Research, MIS Quarterly 31(4): 623–656.<br />

Podsakoff, P.M., MacKenzie, S.B., Jeong-Yeon, L. and Podsakoff, N.P. (2003).<br />

Common Method Biases in Behavioral Research: A critical review of the<br />

literature and recommended remedies, Journal of Applied Psychology 88(5):<br />

879–903.<br />

Podsakoff, P.M. and Organ, D.W. (1986). Self-reports in Organizational<br />

Research: Problems and prospects, Journal of Management 12(4):<br />

531–544.<br />

Powell, W.W. and Colyvas, J.A. (2008). Microfoundations of Institutional<br />

Theory, in R. Greenwood, C. Oliver, K. Sahlin and R. Suddaby<br />

(eds.) Handbook of Organizational Institutionalism, London: Sage,<br />

pp. 276–298.<br />

Qureshi, I. and Compeau, D. (2009). Assessing Between-Group Differences in<br />

Information Systems Research: A comparison of covariance- and<br />

component-based SEM, MIS Quarterly 33(1): 197–214.<br />

Ragu-Nathan, B.S., Apigian, C.H., Ragu-Nathan, T.S. and Tu, Q. (2004). A Path<br />

Analytic Study of the Effect of Top Management Support for Information<br />

Systems Performance, Omega 32(6): 459–471.<br />

Rai, A., Brown, P. and Tang, X. (2009). Organizational Assimilation of<br />

Electronic Procurement Innovations, Journal of Management Information<br />

Systems 26(1): 257–296.<br />

Ringle, C.M., Wende, S. and Will, A. (2005). SmartPLS 2.0 (beta), [WWW<br />

document] http://www.smartpls.de (accessed 28 April 2011).<br />

RMA (2010). Data Quality: Is the glass half empty or half full?, [WWW<br />

document] http://ebiz.rmahq.org/eBusPPRO/Default.aspx?TabId=<br />

55&ProductId=30500 (accessed 28 April 2010).<br />

Rogers, E.M. (1995). Diffusion of Innovations, New York: The Free Press.<br />

Rosenkopf, L. and Abrahamson, E. (1999). Modeling Reputational and<br />

Informational Influences in Threshold Models of Bandwagon Innovation<br />

Diffusion, Computational and Mathematical Organization Theory 5(1):<br />

361–384.<br />

Sambamurthy, V., Bharadwaj, A. and Grover, V. (2003). Shaping Agility<br />

through Digital Options: Reconceptualizing the role of information<br />

technology in contemporary firms, MIS Quarterly 27(2): 237–263.<br />

Sarstedt, M., Henseler, J. and Ringle, C.M. (2011). Multi-Group Analysis in<br />

Partial Least Squares (PLS) Path Modeling: Alternative methods and<br />

empirical results, Advances in International Marketing 22(1): 195–218.<br />

SEC (2008). Bank of America Corporation Form 10-K, [WWW document]<br />

http://www.sec.gov/archives/edgar/data/70858/000119312509041126/<br />

D10K.htm (accessed 28 April 2010).<br />

Sharma, S., Durand, R. and Gur-Arie, O. (1981). Identification and<br />

Analysis of Moderator Variables, Journal of Marketing Research 18(3):<br />

291–299.<br />

Sivo, S.A., Saunders, C., Chang, Q. and Jiang, J.J. (2006). How Low Should You<br />

Go? Low Response Rates and the Validity of Inference in IS Questionnaire<br />

Research, Journal of the Association for Information Systems 7(6):<br />

351–414.<br />

Sternberg, R.J. (2000). Images of Mindfulness, Journal of Social Issues<br />

56(1): 11–26.<br />

Straub, D. (1989). Validating Instruments in MIS Research, MIS Quarterly<br />

13(2): 147–169.<br />

Straub, D., Boudreau, M.-C. and Gefen, D. (2004). Validation Guidelines for IS<br />

Positivist Research, Communications of the Association for Information<br />

Systems 13(24): 380–427.<br />

Strong, P. (2005). Enterprise Grid Computing, ACM Queue 3(6): 50–59.<br />

Sugumaran, V., Tanniru, M. and Storey, V.C. (2008). A Knowledge-Based<br />

Framework for Extracting Components in Agile Systems Development,<br />

Information Technology and Management 9(1): 37–53.<br />

Swanson, E.B. (1994). Information Systems Innovation among Organizations,<br />

Management Science 40(9): 1069–1092.


18<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Swanson, E.B. and Ramiller, N.C. (2004). Innovating Mindfully with<br />

Information Technology, MIS Quarterly 28(4): 553–583.<br />

Tanriverdi, H. (2005). Information Technology Relatedness, Knowledge<br />

Management Capability, and Performance of Multibusiness Firms, MIS<br />

Quarterly 29(2): 311–334.<br />

Teece, D.J. (2007). Explicating Dynamic Capabilities: The nature and<br />

microfoundations of (sustainable) enterprise performance, Strategic<br />

Management Journal 28(13): 1319–1350.<br />

Teece, D.J., Pisano, G. and Shuen, A. (1997). Dynamic Capabilities and Strategic<br />

Management, Strategic Management Journal 18(7): 509–533.<br />

Tenbrunsel, A.E., Galvin, T.L., Neale, M.A. and Bazerman, M.H. (1996).<br />

Cognitions in Organizations, in S.R. Clegg, C. Hardy and W.R. Nord (eds.)<br />

Handbook of Organization Studies, Thousand Oaks, CA: Sage,<br />

pp. 313–337.<br />

Teo, H.H., Wei, K.K. and Benbasat, I. (2003). Predicting Intention to Adopt<br />

Interorganizational Linkages: An institutional perspective, MIS Quarterly<br />

27(1): 19–49.<br />

Thomas, J.B., Clark, S.M. and Gioia, D.A. (1993). Strategic Sensemaking and<br />

Organizational Performance: Linkages among scanning, interpretation,<br />

action, and outcomes, Academy of Management Journal 36(1): 239–270.<br />

Tornatzky, L.G. and Fleischer, M. (1990). The Processes of Technological<br />

Innovation, Massachusetts/Toronto: Lexington Books.<br />

Valorinta, M. (2009). Information Technology and Mindfulness in<br />

Organizations, Industrial and Corporate Change 18(5): 963–997.<br />

Venkatesh, V. and Morris, M.G. (2000). Why Don’t Men Ever Stop to Ask for<br />

Directions? Gender, Social Influence, and Their Role in Technology<br />

Acceptance and Usage Behavior, MIS Quarterly 24(1): 115–139.<br />

Weick, K.E. and Sutcliffe, K.M. (2001). Managing the Unexpected:<br />

Assuring high performance in an age of complexity, San Francisco, CA:<br />

Jossey-Bass.<br />

Weick, K.E. and Sutcliffe, K.M. (2007). Managing the Unexpected: Resilient<br />

performance in an age of uncertainty, San Francisco, CA: Jossey-Bass.<br />

Weick, K.E., Sutcliffe, K.M. and Obstfeld, D. (1999). Organizing for High<br />

Reliability: Processes of collective mindfulness, Research in Organizational<br />

Behavior 21(1): 81–123.<br />

Weinhardt, C., Anandasivam, A., Blau, B., Borissov, N., Meinl, T., Michalik, W.<br />

and Stößer, J. (2009). Cloud Computing: A classification, business models,<br />

and research directions, Business and Information Systems Engineering 1(5):<br />

391–399.<br />

Wold, H. (1982). Soft Modeling: The basic design and some extensions,<br />

in K.G. Jöreskog and H. Wold (eds.) Systems under Indirect<br />

Observation: Causality, structure, prediction, Amsterdam: North-Holland,<br />

pp. 1–54.<br />

Yi, M.Y. and Davis, F.D. (2003). Developing and Validating an Observational<br />

Learning Model of Computer Software Training and Skill Acquisition,<br />

Information Systems Research 14(2): 146–169.<br />

Zhu, K., Dong, S., Xu, S.X. and Kraemer, K.L. (2006a). Innovation Diffusion<br />

in Global Contexts: Determinants of post-adoption digital transformation<br />

of European companies, European Journal of Information Systems 15(6):<br />

601–616.<br />

Zhu, K. and Kraemer, K.L. (2005). Post-Adoption Variations in Usage<br />

and Value of E-Business by Organizations: Cross-country evidence<br />

from the retail industry, Information Systems Research 16(1):<br />

61–84.<br />

Zhu, K., Kraemer, K.L. and Xu, S. (2006b). The Process of Innovation<br />

Assimilation by Firms in Different Countries: A technology diffusion<br />

perspective on e-business, Management Science 52(10): 1557–1576.<br />

Zhu, K., Kraemer, K.L., Xu, S. and Dedrick, J. (2004). Information Technology<br />

Payoff in E-Business Environments: An international perspective on value<br />

creation of e-business in the financial services industry, Journal of<br />

Management Information Systems 21(1): 17–54.<br />

About the author<br />

Martin Wolf is a Research Associate at the Chair of<br />

Business Administration, esp. Information Management<br />

and the E-<strong>Finance</strong> <strong>Lab</strong> at Goethe University, Frankfurt,<br />

Germany. Previously, he earned a diploma in IS and finance<br />

from the University of Mannheim, Germany and the<br />

University of Waterloo, Canada as well as a Ph.D. from<br />

Goethe University, Germany. He gained industry expertise<br />

as part of his assignment in the financial services analytics<br />

division at SAP. His current research focuses on bounded<br />

rationality concepts and their influence in complex<br />

decision-making situations, such as IT innovation assimilation<br />

processes. His articles have been published in<br />

journals such as Business & Information Systems Engineering<br />

and the Australasian Journal of Information Systems as<br />

well as several conference proceedings such as International<br />

Conference on Information Systems and the European<br />

Conference on Information Systems.<br />

Roman Beck is an Assistant <strong>Prof</strong>essor and the E-<strong>Finance</strong><br />

and Services Science Chair at Goethe University in<br />

Frankfurt, Germany, where he also earned his Ph.D. His<br />

research in services science focuses on the role of IT<br />

services sourcing, services management and services<br />

engineering with a special focus on IS outsourcing, social<br />

media and virtualization. From a theoretical perspective, he<br />

is interested in institutional logics of organizations, control<br />

balancing, as well as organizational mindfulness and<br />

awareness. He serves as Senior Editor for the Journal of<br />

Information Technology Theory and Application and has<br />

published over 100 conference papers and articles in<br />

journals such as Communications of the AIS, Information<br />

Technology and People, Scandinavian Journal of Information<br />

Systems, IEEE Transactions on Software Engineering,<br />

Communications of the ACM, Business and Information<br />

Systems Engineering, and others.<br />

Immanuel Pahlke is a Ph.D. candidate at the E-<strong>Finance</strong> and<br />

Services Science Chair of the Institute of Information<br />

Systems at Goethe University, Frankfurt, Germany. He has<br />

a diploma in IS and business economics, which he received<br />

from the Technical University in Darmstadt, Germany. He<br />

also had worked for a management consulting firm in the<br />

financial services sector and currently works as a research<br />

assistant at the E-<strong>Finance</strong> <strong>Lab</strong>. His research interests are in<br />

the areas of end user empowerment and knowledge<br />

management specifically with a focus on enterprise social<br />

media platforms. His articles have been published in<br />

journals such as Business & Information Systems Engineering<br />

and Zeitschrift für Betriebswirtschaftslehre as well as<br />

several conference proceedings such as International<br />

Conference on Information Systems, European Conference<br />

on Information Systems.


Appendix<br />

Table A1 Measurement items<br />

Mindfully resisting the bandwagon M Wolf et al<br />

MP (reflective measures) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Source: Liang et al. (2007)<br />

MP1 Our main competitors who have adopted Grid technology have greatly benefited<br />

MP2 Our main competitors who have adopted Grid technology are favourably perceived<br />

by others in the same industry<br />

MP3 Our main competitors who have adopted Grid technology are favourably perceived<br />

by their suppliers and customers<br />

CP (reflective measures) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Source: Liang et al. (2007)<br />

CP1 The increasing regulatory pressure requires our firm to use Grid technology<br />

CP2 The increasing customer demand requires our firm to use Grid technology<br />

CP3 The competitive conditions require our firm to use Grid technology<br />

NP (reflective measures) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Source: Liang et al. (2007)<br />

NP1 Our firm’s IT services providers have already adopted Grid technology<br />

NP2 Our firm’s business partners have already adopted Grid technology<br />

NP3 The government’s promotion of IT influences our firm to use Grid technology<br />

TMS (reflective measures) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Source: Ragu-Nathan et al. (2004)<br />

TMS1 Top management understands the importance of our Grid infrastructures<br />

TMS2 Top management supports our Grid development projects<br />

TMS3 Top management considers our Grid infrastructures as a strategic resource<br />

TMS4 Top management understands the benefits of our Grid infrastructures<br />

TMS5 Top management keeps the pressure on operating units to use our Grid<br />

infrastructures<br />

ASSM (formative measures) 7-level Guttman (see Table A2)<br />

Source: Rai et al. (2009)<br />

One item for each of the three key business processes (asset management, risk management and new product development<br />

process)<br />

ET (reflective measures) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Sources: Pavlou and El Sawy (2006); Jaworski and Kohli (1993)<br />

ET1 The environment in our industry is continuously changing<br />

ET2 Environmental changes in our industry are very difficult to forecast<br />

ET3 The technology in our industry is changing rapidly<br />

ET4 In our kind of business, customers’ product preferences change a lot over time<br />

ET5 Marketing practices in our product area are constantly changing<br />

ET6 New product introductions are very frequent in our market<br />

ET7 There are many competitors in our market<br />

BPP (formative measures) Extracted latent variable scores (see Table A3)<br />

Sources: Karimi et al. (2007a, b); expert interviews<br />

One item for each of the three key business processes (asset management, risk management and new product development<br />

process)<br />

Control: GIC (reflective measures) 5-point Likert (0%; 25%; 50%; 75%; 100%)<br />

Sources: Zhu and Kraemer (2005); Zhu et al. (2006a, b)<br />

Please estimate the percentage of your firm’s business applications that access y<br />

GIC1 Distributed computing power<br />

GIC2 High-capacity, low-latency networks<br />

GIC3 Service-oriented architectures<br />

GIC4 Computer clusters<br />

GIC5 Distributed databases<br />

19


20<br />

Table A1 Continued<br />

Control: GTI (reflective measures) 5-point Likert (0%; 25%; 50%; 75%; 100%)<br />

Sources: Zhu and Kraemer (2005); Zhu et al. (2006a, b)<br />

Please estimate the percentage of your firm’s business applications that y<br />

GTI1 Are integrated in a service-oriented architecture<br />

GTI2 Are part of an enterprise application integration (EAI) infrastructure<br />

GTI3 Access computing resources from external IT services providers<br />

GTI4 Access cloud computing resources (e.g., Amazon’s EC2)<br />

GTI5 Are executed in a virtual machine<br />

Further controls<br />

Open survey questions and secondary data<br />

(one-item measures)<br />

Sources: Rogers (1995); Fichman (2001)<br />

TIME Years elapsed since the first Grid adoption<br />

SIZE Number of employees (worldwide)<br />

Table A2 Guttman scale for Grid assimilation (ASSM)<br />

Assimilation stage Criteria to enter the assimilation stage Survey items<br />

1. Awareness Key decision makers are aware of Grid<br />

technology<br />

2. Interest The organization is committed to actively learn<br />

more about Grid technology for its PROCESS<br />

3. Evaluation/trial The organization has acquired specific<br />

innovation-related products and has initiated<br />

evaluation or trial for its PROCESS<br />

4. Commitment The organization has committed to use Grid<br />

technology in a significant way for its PROCESS<br />

5. Limited<br />

deployment<br />

6. Partial<br />

deployment<br />

7. General<br />

deployment<br />

Mindfully resisting the bandwagon M Wolf et al<br />

The organization has established a programme of<br />

regular, but limited, use of Grid technology for<br />

part of its PROCESS<br />

The organization has established a programme of<br />

regular, but limited, use of Grid technology for<br />

their PROCESS<br />

The organization has reached a state where Grid<br />

technology is substantially used for its PROCESS<br />

Are you aware of initial or prior Grid-related<br />

activities at site?<br />

Are you aware of plans to use a Grid<br />

environment for PROCESS within the next 12<br />

months?<br />

Is any Grid environment for PROCESS<br />

currently being evaluated or trialled?<br />

Are any Grid application development<br />

projects for PROCESS planned, in progress,<br />

implemented or cancelled?<br />

Are more than 5% but less than 25% of the<br />

business applications for PROCESS running<br />

on a Grid?<br />

Are more than 25% but less than 50% of the<br />

business applications for PROCESS running<br />

on a Grid?<br />

PROCESS ¼ Asset management process/risk management process/new product development process.<br />

Are more than 50% of the business<br />

applications for PROCESS running on a Grid?


Mindfully resisting the bandwagon M Wolf et al<br />

Table A3 Measurement scales for business processes performance second-order construct<br />

BPP (reflective second-order) 7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Sources: Karimi et al. (2007a, b)<br />

Process efficiency (reflective measures)<br />

EFC1 Grid implementation has improved the efficiency of our PROCESS<br />

EFC2 Grid implementation has lowered our costs in PROCESS<br />

EFC3 Grid implementation has decreased the time-to-market of new financial products<br />

due to an improved PROCESS<br />

Process effectiveness (reflective<br />

measures)<br />

EFT1 Grid implementation has improved our effectiveness in PROCESS<br />

EFT2 The functionalities of the Grid adequately meet the requirements of managing<br />

PROCESS<br />

EFT3 Grid implementation has improved our quality of managing PROCESS<br />

Process flexibility (reflective measures)<br />

FLX1 Grid implementation has given us more ways to customize our PROCESS<br />

FLX2 Grid implementation has made our firm more agile due to an improved<br />

PROCESS<br />

FLX3 Grid implementation has made us more adaptive to a changing business<br />

environment due to an improved PROCESS<br />

FLX4 Grid implementation has improved the flexibility of our PROCESS<br />

PROCESS ¼ Asset management process/risk management process/new product development process.<br />

Table A4 Measurement scales for organizational mindfulness score<br />

7-point Likert (1 ¼ strongly disagree; 7 ¼ strongly agree)<br />

Reluctance to simplify<br />

(reflective measures)<br />

RS1 People are encouraged to express different views of the world<br />

RS2 People listen carefully; it is rare that anyone’s view is dismissed<br />

RS3 We appreciate sceptics<br />

Commitment to resilience<br />

(reflective measures)<br />

CR1 People are encouraged to limit any negative consequences so that the firm can continue<br />

operations in case of a mistake<br />

CR2 People are known for their ability to use their knowledge in novel ways<br />

CR3 People have a number of informal contacts that they sometimes use to solve problems<br />

Deference to expertise<br />

(reflective measures)<br />

DE1 If something out of the ordinary happens, people know who has the expertise to respond<br />

DE2 In this organization, the people most qualified to make decisions make them<br />

DE3 People in this organization value expertise and experience over hierarchical rank<br />

Operational sensitivity<br />

(reflective measures)<br />

OS1 Should problems occur, someone with the authority to act is always accessible and available,<br />

especially to people on the front lines<br />

OS2 During an average day, people come into enough contact with each other to build a clear picture<br />

of the current situation<br />

OS3 We have access to resources if unexpected surprises occur<br />

Preoccupation with failure<br />

(reflective measures)<br />

PF1 We treat near misses and errors as information about the health of our system, and try to learn<br />

from them<br />

PF2 People are inclined to report mistakes that could have significant consequences, even if nobody<br />

notices<br />

PF3 People feel free to talk to superiors about problems<br />

21


22<br />

Table A5 Tests for equality of variance and normality among bootstrapping estimations<br />

Relationship Leven’s test (test for equality of variance) Skewness and kurtosis test for normality<br />

OM low (n ¼ 152) OM high (n ¼ 150)<br />

CP-TMS 43.29 (0.00) 20.91 (0.00) 21.47 (0.00)<br />

NP-TMS 71.45 (0.00) 5.87 (0.05) 5.97 (0.05)<br />

MP-TMS 59.21 (0.00) 54.88 (0.00) 13.18 (0.00)<br />

ET-TMS 17.06 (0.00) 0.12 (0.94) 44.57 (0.00)<br />

ET MP-TMS 43.27 (0.00) 6.23 (0.04) 0.42 (0.81)<br />

GIC-ASSM 1.17 (0.28) 21.12 (0.00) 22.83 (0.00)<br />

GTI-ASSM 7.19 (0.01) 4.25 (0.12) 2.43 (0.30)<br />

TIME-ASSM 0.30 (0.58) 20.34 (0.00) 30.36 (0.00)<br />

SIZE-ASSM 32.71 (0.00) 62.27 (0.00) 3.04 (0.22)<br />

TMS-ASSM 50.07 (0.00) 8.56 (0.01) 0.59 (0.75)<br />

GIC-BPP 7.31 (0.01) 43.18 (0.00) 73.47 (0.00)<br />

GTI-BPP 3.53 (0.06) 62.39 (0.00) 29.52 (0.00)<br />

TIME-BPP 18.06 (0.00) 68.06 (0.00) 39.6 (0.00)<br />

SIZE-BPP 11.56 (0.00) 53(0.00) 33.62 (0.00)<br />

ASSM-BPP 1.12 (0.29) 1.17 (0.56) 1.32 (0.52)<br />

ET-BPP 5.75 (0.02) 20.91 (0.00) 22.68 (0.00)<br />

Leven’s test: P-values in parentheses measure the probability of inequality of variance among the two subgroups.<br />

Skewness/Kurtosis test: if the P-value in parentheses is below 0.05, then the null hypothesis (i.e., that the corresponding variable is<br />

normally distributed) is rejected at the 0.05 significance level.<br />

Table A6 Tests for non-response bias based on archival and self-reported data<br />

Indicator Definition and rationale nnon assimilators nassimilators ANOVA<br />

(parametric)<br />

Asset quality<br />

Loan loss reserve/<br />

gross loans<br />

Net charge off/<br />

average gross loans<br />

Reserve for losses arising from risk<br />

exposure, expressed as ratio to total<br />

loans. The higher the ratio. The<br />

poorer will be the quality of the loan<br />

portfolio<br />

Reflects the percentage of today’s<br />

loans in comparison with loans that<br />

have been written off. The lower this<br />

figure the better<br />

Capital structure<br />

Equity/total assets A higher ratio reflects the ability of<br />

the bank to withstand losses<br />

Equity/net loans This ratio measures the equity<br />

cushion available to mitigate losses<br />

in the loan book<br />

Equity/costs of<br />

short-term funding<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Measures the amount of permanent<br />

funding relative to short-term,<br />

potentially volatile funding<br />

Operational performance<br />

Price/earnings ratio This ratio represents investors’<br />

collective opinion regarding a firm’s<br />

future risk and opportunity mix<br />

Earnings per share The portion of a company’s profit<br />

allocated to each outstanding share<br />

of common stock<br />

Kruskal–Wallis<br />

(non-parametric)<br />

39 69 0.21 (0.08) 13.61 (0.02)<br />

39 69 0.33 (0.13) 12.04 (0.06)<br />

39 70 1.49 (0.04) 11.16 (0.08)<br />

39 70 10.72 (0.32) 5.97 (0.34)<br />

39 69 1.49 (0.25) 8.51 (0.18)<br />

15 20 6.67 (0.09) 3.15 (0.23)<br />

15 23 2.45 (0.23) 5.45 (0.11)


Table A6 Continued<br />

Indicator Definition and rationale nnon assimilators nassimilators ANOVA<br />

(parametric)<br />

Net interest margin The higher this ratio, the cheaper<br />

the funding or the higher the margin<br />

the bank is commanding. Higher<br />

margins and profitability are<br />

desirable as long as the asset quality<br />

is being maintained<br />

Non op items and<br />

taxes/average assets<br />

Return on average<br />

assets<br />

Measure of the operating<br />

performance of the bank before tax<br />

and unusual items<br />

Ratio shows how profitable<br />

accompany’s assets are in generating<br />

revenue<br />

Liquidity<br />

Net loans/total assets This liquidity ratio indicates what<br />

percentage of the assets of the bank<br />

are tied up in loans. The higher this<br />

ratio, the less liquid the bank will be<br />

Net loans/costs of<br />

short-term funding<br />

This loans-to-deposit ratio is a<br />

measure of liquidity for which<br />

higher figures denotes lower<br />

liquidity<br />

Kruskal–Wallis<br />

(non-parametric)<br />

39 71 0.21 (0.52) 0.41 (0.05)<br />

39 71 0.13 (0.26) 7.67 (0.22)<br />

39 71 0.01 (0.98) 0.76 (0.91)<br />

39 70 1.74 (0.64) 1,1 (0.86)<br />

39 70 12.29 (0.09) 3.21 (0.61)<br />

All values are calculated based on archival data for 2008, extracted from the BankScope and COMPUSTAT databases.<br />

Definitions of the asset quality, capital structure, operational performance and liquidity indicators were obtained from the BankScope and<br />

COMPUSTAT databases.<br />

ANOVA with Bonferroni test: mean (assimilators) – mean (non-assimilators), with P-values in brackets.<br />

Kruskal–Wallis tests: rank mean (assimilators) – rank mean (non-assimilators), with P-values in brackets.<br />

Table A7 Tests for non-response bias, based on archival and self-reported data<br />

Group comparison: archival data<br />

(n early, n late–between 10 and 30 observations)<br />

Mindfully resisting the bandwagon M Wolf et al<br />

Indicator ANOVA<br />

(parametric)<br />

Kruskal–Wallis<br />

(non-parametric)<br />

Group comparison: self-reported data<br />

(n early ¼ 75, n late ¼ 75)<br />

Question a ANOVA<br />

(parametric)<br />

Kruskal–Wallis<br />

(non-parametric)<br />

Asset quality MP1 0.26 (0.20) 7.94 (0.26)<br />

Loan loss reserve/gross loans 0.52 (0.27) 3.48 (0.39) CP1 0.23 (0.39) 3.94 (0.58)<br />

Net charge off /average gross loans 0.35 (0.32) 3.18 (0.43) NP1 0.16 (0.49) 1.94 (0.78)<br />

Capital structure TMS1 0.28 (0.26) 7.26 (0.30)<br />

Equity/total assets 1.42 (0.24) 4.53 (0.29) ASSM1 0.25 (0.45) 7.70 (0.27)<br />

Equity/net loans 5.25 (0.77) 6.05 (0.15) ET1 0.15 (0.47) 1.86 (0.79)<br />

Operations EFC1 0.11 (0.67) 3.47 (0.56)<br />

Price/earnings ratio 6.57 (0.15) 2.48 (0.32) EFT1 0.29 (0.14) 8.25 (0.18)<br />

Earnings per share 2.05 (0.17) 2.48 (0.33) FLX1 0.18 (0.44) 3.05 (0.61)<br />

Net interest margin 0.45 (0.41) 1.66 (0.69) GIC1 0.08 (0.66) 2.42 (0.72)<br />

Liquidity GTI1 0.07 (0.69) 5,54 (0.43)<br />

Net loans/total assets 3.49 (0.49) 1.81 (0.66) TIME 0.68 (0.20) 12.89 (0.07)<br />

Net loans/costs of short-term Funding 3.48 (0.55) 2.27 (0.59) SIZE 2437 (0.32) 8.68 (0.22)<br />

a Because of space restrictions, we report only the results of the first question (i.e., indicator) for each construct.<br />

ANOVA with Bonferroni test: mean (late responses) – mean (early responses), with P-values in brackets.<br />

Kruskal–Wallis tests: rank mean (late responses) – rank mean (early responses), with P-values in brackets.<br />

Archival data for 2008 were extracted from the BankScope and COMPUSTAT databases.<br />

23


Layer 2:<br />

E-Financial Markets & Market Infrastructures<br />

(<strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber)<br />

� Gomber P., Pujol G., Wranik A. (2012):<br />

Best Execution Implementation and Broker Policies in Fragmented European<br />

Equity Markets<br />

In: International Review of Business Research Papers, Vol. 8, Issue 2, 144-<br />

162.<br />

� Haferkorn M., Lutat M., Zimmermann K. (2012):<br />

The Effect of Single-Stock Circuit Breakers on the Quality of Fragmented<br />

Markets<br />

In: <strong>Finance</strong>Com 2012, Barcelona, Spain.<br />

� Lattemann C., Loos P., Johannes G., Burghof H., Breuer A., Gomber P.,<br />

Krogmann M., Nagel J., Riess R., Riordan R., Zajonz R. (2012):<br />

High Frequency Trading - Costs and Benefits in Securities Trading and its<br />

Necessity of Regulations<br />

In: Business & Information Systems Engineering, Vol. 4, Issue 2, 93-108.<br />

� Siering, M. (2012):<br />

Investigating the Market Impact of Media Sentiment and Investor Attention<br />

In: <strong>Finance</strong>Com 2012; Barcelona, Spain.<br />

� Weber M.C., Wondrak C. (2012):<br />

Measuring the Influence of Project Characteristics on Optimal Software Project<br />

Granularity<br />

In: Proceedings of the 20 th European Conference on Information Systems<br />

(ECIS), Barcelona, Spain.


Best Execution Implementation and Broker Policies in Fragmented<br />

European Equity Markets<br />

Peter Gomber Gregor Pujol Adrian Wranik1<br />

From November 2007, the “Markets in Financial Instruments<br />

Directive” (MiFID) has to be applied by investment firms and<br />

regulated markets in Europe. Investment firms are obliged to make<br />

provisions including processes and IT systems for order routing to<br />

achieve the best possible result for clients in order execution. We<br />

empirically investigate the implementation of best execution<br />

obligations applying a longitudinal analysis of best execution policies<br />

in 2008 and 2009 respectively. The European trading landscape has<br />

changed as competition between established exchanges and new<br />

trading venues has increased significantly. In both studies, 75<br />

policies of German investment firms were analyzed to investigate<br />

how best execution obligations have been implemented and whether<br />

market fragmentation has been considered.<br />

Field of Research: Financial Service and Banking Regulation, European Financial Markets.<br />

1. Introduction<br />

From November 2007, the “Markets in Financial Instruments Directive” (MiFID) has to be<br />

applied by investment firms and regulated markets when providing investment services in<br />

Europe. The central innovations of MiFID are the new classification of trading venues<br />

(regulated markets, multilateral trading facilities (MTF), systematic internalisers), the definition<br />

of “best execution” at a European level and transparency regulations for OTC-trading.<br />

Investment firms are obliged to make adequate provisions including processes and IT systems<br />

for order routing (“best execution arrangements”) to achieve the best possible result and to<br />

disclose sufficient information of the most important measures to clients (“best execution<br />

policies”). Although “best execution” and the associated duties constitute a legal obligation in<br />

the relationship between clients and investment firms, at the economic level this topic also<br />

decisively affects the interface between investment firms and execution venues.<br />

Peter Gomber, Gregor Pujol, Adrian Wranik, Chair of e-<strong>Finance</strong>, E-<strong>Finance</strong>lab, Goethe-University Frankfurt, Germany<br />

email: {gomber | pujol | wranik} @wiwi.uni-frankfurt.de


Best execution is discussed from different perspectives in literature, e.g. how best execution<br />

can be realized and measured. Macey and O’Hara (1996) analyzed legal and economic<br />

aspects of the duty of best execution and recommended that best execution for a particular<br />

trade is best achieved through competition between trading venues. However, McCleskey<br />

(2004) suggested that best execution should be subject to regulation, as investors are not<br />

capable to evaluate execution quality due to limited access to appropriate information. A<br />

number of papers examined costs as a key aspect involved with best execution for a single<br />

trading venue (Roll, 1984; Stoll, 1989) as well as between different markets (Huang & Stoll,<br />

1996; De Jong et al., 1993).<br />

Recent research increasingly focuses on the benefits of technology such as smart order<br />

routing systems to achieve best execution. Foucault and Menkveld (2008) studied the<br />

competition for order flow and concluded that transaction costs could be reduced if market<br />

participants adopted smart order routing. Ende et al. (2009) developed a methodology to<br />

assess advantages of dynamic routing and quantified the economic benefits of smart order<br />

routing technology for European equities.<br />

After the enforcement of MiFID, Hengelbrock and Theissen (2009) studied the market entry of<br />

Turquoise, finding that the new MTF does not provide lower execution costs than primary<br />

markets. Riordan et al. (2010) compared market quality of the London Stock Exchange (LSE)<br />

against a number of MTFs. While quoted spreads are lower on the LSE, implicit transaction<br />

costs measured as effective spreads are on average smaller on Chi-X, BATS, and Turquoise.<br />

With MiFID, the European Regulator particularly intends to harmonize best execution<br />

requirements on a European level. The new regime is characterized by a large diversity of<br />

influencing factors (financial instruments, execution venues) that investment firms have to<br />

consider in order execution processes and IT systems.<br />

Earlier studies among 200 investment firms in Germany revealed that for most German<br />

financial institutions MiFID is more of a regulatory burden than a chance to leverage<br />

competitive potentials (Gomber et al., 2007). However, 32% of the investment firms answered<br />

that competitive differentiation can be achieved through the design of best execution policies<br />

and considered this aspect as having the highest chances of all services connected with<br />

MiFID.<br />

Thus, the purpose of this paper is to analyze the evolution of the implementation of best<br />

execution processes and applied IT systems as documented in best execution policies after<br />

MiFID entered into force. Different policies of German investment firms are analyzed and<br />

compared over time. The study also checks how far the institutions' earlier assessment of the<br />

competitive potential of order execution is reflected in best execution policies.<br />

The motivation for a longitudinal analysis, i.e. the collection and analysis of the best execution<br />

policies both in 2008 (right after the initial implementation of MiFID) and also in 2009 is twofold:<br />

First, MiFID mandates a review of the investment firms’ policies at least on an annual<br />

basis or whenever major changes occur, i.e. we can assure that investment firms investigated<br />

and adapted their policies between the two points of analysis. Second, the European trading<br />

landscape has experienced an intensified competition after the enforcement of the MiFID that<br />

unfolded its impact on the market shares among trading venues between 2008 and 2009,<br />

especially with the advent of new trading venues (MTF) offering serious alternatives for best<br />

execution.


The next chapter provides some background on developments of European financial markets.<br />

Chapter 3 presents the research questions and sample data. The main results are reported in<br />

chapter 4. Chapter 5 concludes.<br />

2. MiFID impact on European trading venues<br />

MiFID has triggered a new competitive environment for equity trading and services. With the<br />

proliferation of new MTFs such as Chi-X, Turquoise or BATS Europe, the number of trading<br />

venues has substantially increased offering a wider range of trading options.<br />

Two key trends can be observed: First, order flow is increasingly directed away from<br />

incumbent exchanges towards new MTFs. Second, cost structures for trading and post-trading<br />

activities in the European marketplace have changed significantly.<br />

Due to MiFID, market shares have been shaken up and trading volumes have recognizably<br />

shifted among execution venues. Figure 1 illustrates the changes for three important European<br />

equity indices over time. While prior to MiFID domestic markets held market shares close to<br />

100%, new trading platforms gradually have gained importance, e.g. for FTSE-100 the LSE’s<br />

market share accounts for just 50%.<br />

Figure 1: Market shares of major European indices over time (Fidessa, 2010)<br />

DAX30<br />

CAC40<br />

FTSE100<br />

July 2008<br />

1st July 2008<br />

1 MiFID study<br />

st MiFID study<br />

Xetra<br />

Chi-X<br />

Paris<br />

Chi-X<br />

LSE<br />

Chi-X<br />

FTSE100<br />

4%<br />

4%<br />

July 2009<br />

2nd July 2009<br />

2 MiFID study<br />

nd MiFID study<br />

Xetra<br />

Chi-X<br />

Turquoise<br />

Bats Europe<br />

Others<br />

Paris<br />

Chi-X<br />

Amsterdam<br />

Bats Europe<br />

Turquoise<br />

Others<br />

LSE<br />

Chi-X<br />

Turquoise<br />

Bats Europe<br />

Others<br />

April 2010<br />

Xetra<br />

Chi-X<br />

Bats Europe<br />

Turquoise<br />

Others<br />

Paris<br />

Chi-X<br />

Amsterdam<br />

Bats Europe<br />

Turquoise<br />

Others<br />

LSE<br />

Chi-X<br />

Bats Europe<br />

Turquoise<br />

Nasdaq Europe<br />

In 2009, the European Commission commissioned a study (Oxera, 2009) to monitor fee<br />

developments across 18 European markets. One key result addresses the explicit costs for<br />

trading services that on average have decreased along the entire trading value chain.<br />

Others<br />

time


Figure 2: Change in explicit costs over time (2006-2008) (Oxera, 2009)<br />

80<br />

60<br />

40<br />

20<br />

0<br />

-20<br />

-40<br />

-60<br />

-80<br />

-100<br />

Trading platforms Central Counterparties Central Securities<br />

(CCPs) Depositories (CSD)<br />

�� Major financial centres: Germany, France, UK, Italy, Switzerland and Spain<br />

�� Secondary financial centres: Belgium, Luxemburg, Netherland, Norway, Poland and Sweden<br />

▲ Other financial centres: Denmark, Greece, Ireland, Austria, Portugal und Tschech Republic<br />

Figure 2 highlights changes for different groups of European markets from 2006 to 2008.<br />

Across all groups, average fees charged by trading platforms as well as central counterparty<br />

clearing fees have significantly reduced. Fees charged by Central Securities Depositories do<br />

not reveal a systematic trend.<br />

3. Research questions and data<br />

MiFID does not prescribe how investment firms have to implement best execution obligations.<br />

On the one hand they need to establish adequate internal provisions for business processes<br />

and IT (best execution arrangements), on the other hand sufficient information regarding these<br />

arrangements has to be communicated to clients (best execution policies). As only latter<br />

information is publicly disclosed, the subsequent analysis focuses on publicly available data<br />

(best execution policies).<br />

The new European concept of best execution in securities trading and its practical<br />

implementation raise a number of important research questions: (i) How do firms implement<br />

best execution obligations from a process and IT perspective and which criteria are used for<br />

order routing? (ii) Do best execution policies implement the mandatory legal requirements of<br />

MiFID and do they reflect increased competition among trading venues over time? (iii) Which


execution venues are preferred and how does this preference evolve due to the changed<br />

competitive landscape?<br />

The paper takes four steps to address these research questions:<br />

First, it analyses which option the investment firms selects to implement best execution<br />

requirements. It is investigated whether a static rule framework applying historical data only is<br />

used to determine the venues for order execution or whether a dynamic order routing process<br />

is applied that relies on sophisticated order routing software and real time market data to<br />

determine which execution venue is able to provide the best result at the time of order<br />

submission. This analysis of static versus dynamic best execution policies is directly linked to<br />

the criteria an investment firm takes into account for the selection of an execution venue and<br />

therefore these criteria are analyzed in this context. Second, the best execution policies are<br />

investigated as a whole and compared with each other regarding the level of fulfillment of the<br />

mandatory legal requirements over time. Third, the policies were analyzed with regard to the<br />

existence of a ranking for specific client categories or classes of instruments.<br />

As MiFID had to be transformed into individual national legislations by all European member<br />

states, a consistent view can only be achieved for one member state. For both studies<br />

Germany was selected as it represents the largest European economy. For the purpose of<br />

investigating the changes of the best execution implementation and the reaction of investment<br />

firms on the recent developments in the European trading landscape both studies use an<br />

identical sample of firms’ best execution policies in Q2/2008 and Q3/2009 respectively.<br />

Various channels were used for obtaining the policies, e.g. by data collection over the Internet,<br />

by email or telephone contacts.<br />

The study is based on the 100 largest German financial institutions in terms of total assets in<br />

2006 (Karsch, 2007) and the 15 largest online brokers according to number of security<br />

accounts (Kundisch & Holtmann, 2008). The list of the 100 largest institutions was adjusted by<br />

removing the companies which do not provide investment services reducing the number to 63.<br />

Since in 2008, three best execution policies were not made available, the final sample covers<br />

75 best execution policies.<br />

In the following, results of the data analysis in 2009 are presented and compared to the<br />

findings in 2008. In subsequent figures and tables, the results of the data analysis in 2009 are<br />

marked in bold, the comparative values of 2008 are indicated in parentheses. While in the text,<br />

percentage values always refer to the total sample size (75 policies), absolute figures are used<br />

to explain findings of a sub-sample.<br />

4. Results<br />

4.1 Implementation of Best Execution Processes<br />

Best execution policies are a major part of the provisions which firms must make in order to<br />

ensure that they can regularly execute orders in the best possible manner. The most important<br />

legal requirements of MiFID implementation in Germany are specified in the German<br />

Securities Trading Act (WpHG) i and the Ordinance Specifying Rules of Conduct and<br />

Organisation Requirements for Investment Firms (WpDVerOV). ii


Although MiFID lacks guidance on how to implement the best execution obligation, two basic<br />

concepts can be distinguished: Firms may apply a static approach, i.e. the decision to route<br />

client orders to a particular venue is based on a pre-defined rule framework taking into account<br />

different criteria such as client category (retail or professional client), order types and sizes and<br />

classes of instruments (shares, bonds, derivatives). The Committee of European Securities<br />

Regulators (CESR) suggests a minimum level of differentiation (CESR, 2007) by distinguishing<br />

between different client categories and classes of instruments. The result of such a rule<br />

framework typically leads to one particular venue that provides the best possible result for a<br />

particular client category, class of financial instrument or a combination of both based on<br />

historical data.<br />

Alternatively, the appropriate execution venue is selected by applying a dynamic approach<br />

which is an implementation that exceeds MiFID minimum legal requirements: Each order is<br />

treated in an individual manner considering real time market data for the order routing<br />

decision. Such provisions enable a real-time evaluation and support a dynamic allocation of<br />

the individual order to the venue offering the best conditions at the time of order entry.<br />

Considering the increased competition in Europe such a real time comparison allows an<br />

optimized selection between venues improving execution quality but it obviously leads to<br />

higher costs of implementation.<br />

Figure 3: Static vs. dynamic best execution approach<br />

Static<br />

approach<br />

Exchange Exchange Exchange Exchange Exchange Exchange Exchange AA<br />

AA<br />

A AA<br />

A<br />

Buy Buy 100 100 100 Daimler<br />

Daimler<br />

Daimler<br />

Client Client order order order order order order<br />

order order order order order<br />

- Instrument: Daimler<br />

- Index: DAX<br />

- Quantity: 100<br />

- Buy/Sell: Buy<br />

Pre Pre-defined<br />

Pre Pre-defined<br />

Pre<br />

Pre defined defined rule rule<br />

rule framework framework framework framework<br />

framework<br />

framework<br />

Domestic Domestic/ Domestic Domestic/ Domestic Instrument<br />

Instrument Execution<br />

Execution<br />

Category<br />

Category<br />

abroad<br />

abroad class<br />

class class class venue<br />

venue<br />

Shares domestic DAX A<br />

Shares domestic MDAX B<br />

… …<br />

Buy 100 @ 40,80 €<br />

Exchange B<br />

Buy 100 @ 40,70 €<br />

Buy 100 @ 40,70 €<br />

Routing Routing routines<br />

routines<br />

routines<br />

Smart Smart Order Order Order Order Order Order<br />

Order<br />

Routing<br />

Routing<br />

System<br />

System<br />

MTF A<br />

real real-time real real-time real real real real time time market market<br />

market data<br />

data<br />

data<br />

Dynamic<br />

approach<br />

Buy 100 @ 40,75 €<br />

Figure 3 illustrates both concepts. While in the static approach (dashed box) client orders are<br />

routed to the best execution venue (here: Exchange A) based on a pre-defined rule framework,<br />

the dynamic approach (solid box) applies real-time market data in the decision process. Based


on routing routines of the smart order routing system the best venue (here: Exchange B) is<br />

detected at order entry and subsequently the order is submitted to that venue for execution.<br />

Relevant Criteria and Implementation of best execution processes<br />

For both concepts, the relevant criteria for achieving the best possible result are derived<br />

directly from §33a (2) WpHG, in accordance to which “in particular the prices of the financial<br />

instruments, the costs involved in order execution, the speed, the probability of execution and<br />

the processing of the order, as well as the scope and type of order” must be taken into account<br />

as criteria. It was checked whether the relevant criteria had been weighted in the best<br />

execution policies (Table 1).<br />

No weighting can be recognized in 8.0% (2008: 10.7%) (absolute: 6 (2008: 8)) of the policies.<br />

In two cases no details are provided, in the other four cases the criteria have not been<br />

prioritized. Weighting of relevant criteria can be recognized in 92.0% (89.3%) of policies<br />

(absolute: 69 (67)). In 13 (12) policies percentage values (e.g. price: 80%, external costs: 20%)<br />

are named for individual criteria; in 56 (55) policies a ranking, e.g. price priority over speed,<br />

can be observed. In both studies only one policy provides a recognizable ranking of criteria<br />

and uses real-time market data to identify the execution venue providing the best result for the<br />

individual order.<br />

Table 1: Recognizability of the weighting of the relevant criteria and implementation of best<br />

execution processes<br />

Number of policies evaluated 75 (75)<br />

No details 2 (2)<br />

No ranking of criteria recognizable 4 (6)<br />

Ranking of the criteria recognizable and<br />

dynamic order execution (real time)<br />

1 (1)<br />

Ranking of the criteria recognizable 55 (54)<br />

Percentage weighting of the criteria 13 (12)<br />

No recognizable<br />

weighting<br />

Recognizable<br />

weighting<br />

6 (8)<br />

69 (67)<br />

Concerning the different implementation options for best execution processes and systems,<br />

Table 1 shows an important result: For both data analyses (2008 and 2009) it turns out that an<br />

overwhelming majority of firms prefers a static implementation approach for best execution and<br />

does not utilize the competitive potential stated in recent studies.<br />

4.2 Legal Requirements for the Best Execution Policies<br />

According to §33a (6) No. 1 WpHG an investment firm must inform “its clients of its best<br />

execution policy before providing investment services for the first time and obtain the clients'<br />

acceptance of its policy.”<br />

Specific minimum requirements with regard to the scope and design of the contents are linked<br />

to this obligation specified in §33a (5) and (6) WpHG in conjunction with §11 (4) WpDVerOV:<br />

investment firms must show how they aim to achieve the best possible result for order<br />

execution in various categories of financial instruments and the crucial factors for selecting a


venue must be named, and finally a list containing at least those venues must be provided<br />

which can be considered as consistently achieving the best possible results.<br />

Content of the law<br />

Description of the weighting defined for<br />

the relevant criteria<br />

or<br />

description of the method which is used<br />

for weighting<br />

Details of the various execution venues<br />

for each category of financial instrument<br />

Details of the crucial factors for selecting<br />

an execution venue<br />

List of the principal execution venues<br />

which consistently achieve the best<br />

possible result when executing client<br />

orders<br />

Information for private clients in the case<br />

of orders with client instructions<br />

Releasing the investment firms from the<br />

obligation to execute an order in<br />

accordance with the best execution policy<br />

Figure 4: Legal requirements<br />

Policies: 2009: 75 (2008): 75<br />

No details<br />

13 (12)<br />

% - value<br />

56 (55) 4(6) 2 (2)<br />

ranking<br />

method<br />

Yes No<br />

71 (71) 4 (4)<br />

73 (73)<br />

65 (64)<br />

10 (11)<br />

2 (2)<br />

71 (71) 4 (4)<br />

0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100%<br />

Figure 4 shows the results for the legal requirements which. It displays the requirements and<br />

the levels of fulfilment.<br />

Description of the Weighting Implemented or Description of the Method<br />

§ 33a (6) No. 1 WpHG in conjunction with §11 (4) No. 1 WpDVerOV requires that policies must<br />

contain either a “description of the weighting implemented for the relevant criteria to achieve<br />

the best possible result” or “a description of the method which is used for this weighting.” In the<br />

analysis a distinction is made between three examination criteria (percentage, ranking,<br />

method). Policies which contain at least information on a particular procedure are<br />

acknowledged to have a method (e.g. “the bank assumes that their clients would like to<br />

achieve the best price”). In the policies both analysed in 2009 and 2008, this applies for nearly<br />

all investment firms (97.3%) except for two which did not provide any details.<br />

It was investigated whether the weighting defined was expressed as a percentage value or by<br />

a particular ranking. In 17.3% (16.0%) of the best execution policies concrete percentage<br />

values are specified for individual criteria (e.g. price: 80%; external costs: 20%), in 74.4%<br />

(73.3%) a ranking, e.g. price has priority over speed, is provided.<br />

Details of Execution Venues for each Category of Financial Instrument<br />

§33a (5) No. 1 WpHG (first half sentence) requires “details of the various execution venues<br />

with regard to each category of financial instrument.” This provision specifies that a category<br />

must be assigned to the venue(s) when a policy is created. The following variants of policies<br />

exist: either precisely one venue is specified for one category or several venues are specified


for one category or several venues are specified for multiple categories simultaneously. In<br />

94.6% of the policies an assignment of category to venue can be recognized.<br />

Details of the Crucial Factors for Selecting an Execution Venue<br />

In addition in accordance with §33a (5) No. 1 WpHG, “the crucial factors for selecting an<br />

execution venue” must also be specified. In the analysis, information was taken into account<br />

whether documents reveal that the firms used various factors for the evaluation (e.g. “in<br />

particular the recognizable factors price and costs which arise through execution at an<br />

execution venue are used for the evaluation.”). Like in 2008, nearly all policies, i.e. 97.3%,<br />

provided comprehensive details.<br />

List of the Principal Execution Venues<br />

A further obligation derives from §33a (5) No. 2 WpHG in conjunction with §11 (4) No. 2<br />

WpDVerOV according to which a “list of the principal execution venues [�] at which the<br />

investment firms can consistently achieve the best possible results when executing client<br />

orders” must be contained in the policies. In 86.7 % (85.3%) of the policies examined in 2009<br />

this list is provided either as a text list or as a table in the appendix.<br />

Information for Private Clients about Execution in accordance with Instructions<br />

Finally, in accordance with §33a (6) No. 2 WpHG in conjunction with §11 (4) No. 3 WpDVerOV<br />

firms must inform “private clients expressly that when instructed by the client the investment<br />

firm will execute the order according to these client instructions and will therefore not be<br />

obliged to execute the order in accordance with its best execution policy to achieve the best<br />

possible result.” Both in the 2009 and the 2008 analysis, this information is clearly emphasized<br />

in most execution policies, i.e. in 94.7%.<br />

4.3 Analysis of the Execution Venues Specified<br />

Finally, details regarding the ranking of the venues were examined. Obviously, from a<br />

competitive view, this ranking is highly relevant for execution venues. Therefore, it is of interest<br />

how policies list and rank the different venues. Table 2 shows the distribution of the<br />

nominations for various securities groups, classified according to whether they can be traded<br />

domestically or abroad. It is noticeable that firms primarily prefer abstract and summarizing<br />

descriptions to document their choice of a venue (e.g. domestic execution venue, foreign<br />

exchange) instead of naming specific venues. In addition to the abstract details relating to the<br />

venues, the table also includes special cases, such as forwarding, fixed-price business and<br />

also obligations to provide instructions. A concrete venue is named only in every fourth policy<br />

(26.6%; 2008: 24%).<br />

Table 2: Execution venues for different securities groups


Tradable<br />

Tradable<br />

domestically<br />

domestically<br />

Shares Shares Bonds<br />

Bonds<br />

Certified<br />

Certified<br />

derivatives<br />

derivatives<br />

Investment<br />

Investment<br />

shares<br />

shares<br />

ETF<br />

ETF<br />

Other<br />

Other<br />

securities<br />

securities<br />

Other<br />

Other<br />

financial<br />

financial<br />

instruments<br />

instruments<br />

Domestic<br />

execution venue<br />

42 (43) 42 (43) 41 (42) 5 (5) 4 (4) 11 (10) 4 (4)<br />

Domestic floor<br />

trading system<br />

2 (4) 4 (5) 4 (5) 1 (1) 4 (4) 0 (0) 0 (0)<br />

Domestic<br />

exchange<br />

7 (6) 11 (10) 12 (11) 6 (6) 9 (9) 5 (5) 6 (6)<br />

Domestic home<br />

exchange<br />

0 (0) 3 (2) 2 (2) 0 (0) 0 (0) 0 (1) 1 (1)<br />

Others (e.g.<br />

forwarding)<br />

1 (1) 3 (4) 4 (5) 1 (1) 0 (0) 1 (1) 1 (1)<br />

Fixed-price<br />

business<br />

1 (1) 2 (3) 0 (1) 0 (0) 0 (0) 2 (3) 0 (0)<br />

Instructions 0 (0) 1 (1) 1 (1) 46 (46) 2 (2) 34 (34) 35 (35)<br />

No details 2 (2) 1 (1) 3 (3) 13 (13) 49 (48) 19 (19) 9 (9)<br />

Not possible 0 0 (0) (0) 0 0 (0) (0) 0 0 (0) (0) 2 2 (2) (2) 0 0 (1) (1) 0 0 (0) (0) 1 1 (1)<br />

(1)<br />

Specific<br />

execution venue<br />

20 (18) 8 (6) 8 (5) 1 (1) 7 (7) 3 (3) 18* (18)<br />

Total 75 (75) 75 (75) 75 (75) 75 (75) 75 (75) 75 (75) 75 (75)<br />

Tradable<br />

abroad<br />

Shares Bonds<br />

Certified<br />

derivatives<br />

Investment<br />

shares<br />

ETF<br />

Other<br />

securities<br />

Other<br />

financial<br />

instruments<br />

Foreign<br />

execution venue<br />

2 (2) 5 (5) 3 (3) 1 (1) 1 (1) 2 (2) 13 (12)<br />

Foreign exchange 27 (27) 18 (17) 17 (17) 0 (0) 0 (0) 11 (10) 1 (1)<br />

Others (e.g.<br />

forwarding)<br />

4 (4) 8 (6) 8 (8) 2 (2) 2 (2) 2 (2) 2 (2)<br />

Fixed-price<br />

business<br />

1 (2) 1 (1) 0 (0) 0 (0) 0 (0) 0 (1) 1 (2)<br />

Instructions 36 (35) 34 (36) 34 (34) 45 (46) 45 (46) 34 (34) 34 (34)<br />

No details 4 (4) 9 (10) 13 (13) 25 (24) 25 (24) 26 (26) 23 (23)<br />

Not possible 0 (0) 0 (0) 0 (0) 2 (2) 2 (2) 0 (0) 1 (1)<br />

Specific<br />

execution venue<br />

1 (1) 0 (0) 0 (0) 0 (0) 0 (0) 0 (0) 0 (0)<br />

Total 75 (75) 75 (75) 75 (75) 75 (75) 75 (75) 75 (75) 75 (75)<br />

* 16 (16) nominations here for Eurex<br />

For the categories of financial instruments which can be traded domestically, the abstract label<br />

“domestic execution venue” is chosen most frequently for the first three securities groups<br />

(shares, bonds, certified derivatives). For investment shares, other securities and other<br />

instruments orders are mostly accepted with explicit instructions only. Firms most frequently<br />

provide - with 20 (18) nominations - a specific execution venue for the securities group shares.<br />

As in 2008, the securities group other financial instruments was named 18 times, thereof 16<br />

nominations for Eurex. The term “foreign exchange” is frequently used for financial instruments<br />

which can be traded abroad; with one exception, no specific details are provided here. It is<br />

noticeable that for all securities groups order executions in most cases are only possible with<br />

instructions or no details about the execution venues are provided at all.<br />

Due to the few specific details about the execution venues and their ranking, a more detailed<br />

analysis was only performed for the securities group shares.<br />

In Table 4, 20 (18) policies are listed which at least named one specific venue for this<br />

securities group and documented a recognizable ranking.<br />

Table 3: Segmenting of the securities group shares


Segment Segment DAX DAX 30<br />

30<br />

Other Other DAX<br />

DAX<br />

(MDAX, (MDAX, TECDAX,<br />

TECDAX,<br />

SDAX)<br />

SDAX)<br />

EUROSTOXX EUROSTOXX 50,<br />

50,<br />

DJ DJ STOXX STOXX 40,<br />

40,<br />

NASDAQ NASDAQ 100<br />

100<br />

Other Other domestic domestic shares<br />

shares<br />

Execution venue Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2<br />

Xetra-Best 4 (4) 4 (4) 0 (0) 4 (4) 4 (4) 0 (0) 3 (3) 3 (3) 0 (0) 1 (1) 1 (1) 0 (0)<br />

Xetra 11 (14) 6 (8) 5 (5) 13 (14) 6 (8) 7 (5) 8 (10) 4 (5) 4 (4) 9 (11) 5 (6) 4 (4)<br />

Berlin 1 (2) 0 (0) 1 (1) 1 (2) 0 (0) 1 (1) 2 (2) 1 (1) 1 (0) 1 (2) 0 (0) 1 (1)<br />

Düsseldorf 3 (2) 0 (0) 3 (1) 1 (2) 0 (0) 1 (1) 2 (2) 1 (1) 1 (0) 1 (2) 0 (0) 1 (1)<br />

Frankfurt 5 (4) 2 (0) 3 (3) 5 (4) 2 (0) 3 (3) 5 (4) 2 (1) 3 (2) 6 (4) 3 (2) 3 (2)<br />

Hamburg 3 (2) 3 (1) 0 (0) 1 (2) 1 (1) 0 (0) 3 (2) 3 (2) 0 (0) 1 (2) 1 (2) 0 (0)<br />

Hannover 2 (1) 2 (0) 0 (0) 0 (1) 0 (0) 0 (0) 0 (1) 0 (1) 0 (0) 0 (1) 0 (1) 0 (0)<br />

Munich 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2)<br />

Stuttgart 1 (2) 0 (0) 1 (1) 3 (2) 2 (0) 1 (1) 4 (2) 1 (1) 3 (0) 3 (2) 2 (0) 1 (1)<br />

Tradegate 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0)<br />

OTC 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0)<br />

Domestic floor<br />

trading system<br />

3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 4 (4) 1 (1) 3 (3) 5 (4) 2 (2) 3 (2)<br />

Domestic home<br />

exchange<br />

4 (2) 0 (0) 4 (2) 4 (2) 0 (0) 4 (2) 3 (2) 0 (0) 3 (2) 3 (1) 0 (0) 3 (2)<br />

Fixed-price<br />

business<br />

3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0)<br />

No details 3 Execution venue Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2 Freq. Rank 1 Rank 2<br />

Xetra-Best 4 (4) 4 (4) 0 (0) 4 (4) 4 (4) 0 (0) 3 (3) 3 (3) 0 (0) 1 (1) 1 (1) 0 (0)<br />

Xetra 11 (14) 6 (8) 5 (5) 13 (14) 6 (8) 7 (5) 8 (10) 4 (5) 4 (4) 9 (11) 5 (6) 4 (4)<br />

Berlin 1 (2) 0 (0) 1 (1) 1 (2) 0 (0) 1 (1) 2 (2) 1 (1) 1 (0) 1 (2) 0 (0) 1 (1)<br />

Düsseldorf 3 (2) 0 (0) 3 (1) 1 (2) 0 (0) 1 (1) 2 (2) 1 (1) 1 (0) 1 (2) 0 (0) 1 (1)<br />

Frankfurt 5 (4) 2 (0) 3 (3) 5 (4) 2 (0) 3 (3) 5 (4) 2 (1) 3 (2) 6 (4) 3 (2) 3 (2)<br />

Hamburg 3 (2) 3 (1) 0 (0) 1 (2) 1 (1) 0 (0) 3 (2) 3 (2) 0 (0) 1 (2) 1 (2) 0 (0)<br />

Hannover 2 (1) 2 (0) 0 (0) 0 (1) 0 (0) 0 (0) 0 (1) 0 (1) 0 (0) 0 (1) 0 (1) 0 (0)<br />

Munich 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2) 1 (2) 0 (0) 1 (2)<br />

Stuttgart 1 (2) 0 (0) 1 (1) 3 (2) 2 (0) 1 (1) 4 (2) 1 (1) 3 (0) 3 (2) 2 (0) 1 (1)<br />

Tradegate 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0)<br />

OTC 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0) 1 (1) 1 (1) 0 (0)<br />

Domestic floor<br />

trading system<br />

3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 4 (4) 1 (1) 3 (3) 5 (4) 2 (2) 3 (2)<br />

Domestic home<br />

exchange<br />

4 (2) 0 (0) 4 (2) 4 (2) 0 (0) 4 (2) 3 (2) 0 (0) 3 (2) 3 (1) 0 (0) 3 (2)<br />

Fixed-price<br />

business<br />

3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0) 3 (3) 3 (3) 0 (0)<br />

No details 2 (3) 0 (0) 2 (3) 2 (3) 0 (0) 2 (3) 6 (6) 2 (1) 4 (5) 7 (6) 1 (0) 6 (6)<br />

Total 22¹ (18) 23² (21)² 20 (18) 23² (21)² 23² (22)¹² 23² (18) 20 (19)¹ 23² (21)²<br />

1 The execution venues Hamburg / Hanover were counted twice in Rank 1<br />

2 The execution venues Frankfurt/ Stuttgart/ Düsseldorf/ Berlin were counted four times in Rank 1 or 2<br />

3 Some best execution policies do not contain specific details for every segment: for example, in the segment<br />

„EUROSTOXX 50“ the two (one) nomination(s) for Rank 1 mean(s) that two (one) investment firm provided no details<br />

about order execution in this segment, but did document a ranking for the other segments (e.g. DAX 30). As regards the<br />

four (five) nominations for Rank 2: the best execution policies only envisage one specific execution venue being ranked<br />

first, but no further information is providedabout alternative execution venues in the following ranks.<br />

3 2 (3) 0 (0) 2 (3) 2 (3) 0 (0) 2 (3) 6 (6) 2 (1) 4 (5) 7 (6) 1 (0) 6 (6)<br />

Total 22¹ (18) 23² (21)² 20 (18) 23² (21)² 23² (22)¹² 23² (18) 20 (19)¹ 23² (21)²<br />

1 The execution venues Hamburg / Hanover were counted twice in Rank 1<br />

2 The execution venues Frankfurt/ Stuttgart/ Düsseldorf/ Berlin were counted four times in Rank 1 or 2<br />

3 Some best execution policies do not contain specific details for every segment: for example, in the segment<br />

„EUROSTOXX 50“ the two (one) nomination(s) for Rank 1 mean(s) that two (one) investment firm provided no details<br />

about order execution in this segment, but did document a ranking for the other segments (e.g. DAX 30). As regards the<br />

four (five) nominations for Rank 2: the best execution policies only envisage one specific execution venue being ranked<br />

first, but no further information is providedabout alternative execution venues in the following ranks.<br />

Specifically, the segments iii DAX30, other DAX, EUROSTOXX 50, DJ STOXX 40, NASDAQ<br />

100 and other domestic shares were analyzed.<br />

The venue mentioned most frequently in all segments is Xetra. In some policies precisely one<br />

venue is prioritized and occupies rank 1; in these cases no specific nomination for rank 2<br />

exists - a “domestic floor trading system” or “domestic home exchange” was ranked second.<br />

This is why these abstractly mentioned execution venues are also contained in the table. As<br />

execution venues of equal rank are also named twice or four times in the policies, these<br />

multiple nominations result in a value greater than 20 (18) in the bottom line. In three execution<br />

policies order execution as fixed-price business is placed in Rank 1. iv It is noticeable that - like<br />

in 2008 - regional exchanges are rarely placed in rank 1.<br />

The analysis reveals a remarkable finding: The substantial changes of the European trading<br />

landscape triggered by MiFID are not reflected at all in any best execution policy analyzed in<br />

the data of 2009. This becomes apparent in the minor changes identified in Table 4 and the<br />

fact that no new venue (MTF) is considered in any best execution policy.<br />

5. Summary and Outlook<br />

This paper compares the implementation of European best execution obligations in Germany<br />

based on analyses of best execution policies in Q2/2008 and Q3/2009. The analysis aims at<br />

identifying changes in the firm’s policies as a result of the mandatory annual reviews and<br />

investigates whether recent developments in the trading landscape are reflected.<br />

We analysed 75 policies of the 100 largest financial institutions and of the 15 largest online<br />

brokers in Germany. 59% (44) policies remain unchanged compared to 2008. 41% (31) of the<br />

firms provided an updated version of their policy.<br />

The key message of the study is that the results of the two data analyses are largely in line<br />

and no substantial changes can be identified although the European securities markets were


affected by major changes concerning the competitive landscape and the number of available<br />

venues. In both years, the policies show that the minimum legal requirements have<br />

recognizably been implemented. Like in 2008, in 2009 only one policy mentions the use of a<br />

dynamic order execution approach, i.e. applying real time market data to achieve best<br />

execution. Furthermore, a significant heterogeneity can be recognized between the policies:<br />

some are extremely comprehensive and describe the selected procedure in great detail, while<br />

others are limited to minimum details and are not very meaningful for clients.<br />

This also applies for the details about the venues: Only in 26.6% (24.0 %), i.e. in approximately<br />

every fourth policy, the venues are named specifically or a ranking is provided.<br />

In earlier studies (Gomber et al., 2007), best execution principles are most frequently named<br />

as key competitive factor. However, in summarizing it must be noted that the use of policies as<br />

competitive instrument cannot be recognized in a large majority of German financial<br />

institutions. This is surprising given the intensified competition between incumbent and new<br />

trading venues resulting in market shares shifts and cost reductions. These changes are not<br />

reflected in any policy so far. Orders are still primarily executed on domestic markets or (if only<br />

tradable abroad) on the foreign exchange respectively. Furthermore, MTFs are not considered<br />

at all in any best execution policy.<br />

As policy implication of our research and as an input for the currently ongoing MiFID review of<br />

the EU Commission (with a final new regulation scheduled for 2012), it must be noted that<br />

these changes might possibly be implemented by firms internally but they are not listed<br />

explicitly in the policies. Nevertheless, from the client's viewpoint it is desirable that these<br />

analyses and their concrete results should be communicated in a more transparent form so<br />

that the policies can function not just as regulatory necessity but also as driver of competition<br />

between firms and also between venues. For future research, both analyses serve as a basis<br />

for benchmarking best execution policies over longer time periods, e.g. to monitor whether<br />

firms change their processes and systems from a static towards a dynamic approach for order<br />

routing and execution. Also, studies regarding the client benefits of offering access to the new<br />

MTFs or analyses regarding the MiFID perception from a client perspective and its potential<br />

implications in terms of changes in reputation and performance of investment firms represent<br />

challenging research topics.<br />

6. References<br />

CESR 2007. Best Execution under MiFID, Questions & Answers, Ref CEST/07-320.<br />

De Jong, F., Nijman, T. and Roell, A. 1993. A comparison of the cost of trading French shares<br />

on the Paris Bourse and on Seaq International, European Economic Review, 39, 7, pp.<br />

1277-1301.<br />

Ende, B., Gomber, P. and Lutat, M. 2009. Smart Order Routing Technology in the New<br />

European Equity Trading Landscape, Software Services for e-Business and e-Society,<br />

9th IFIP WG 6.1 Conference, I3E 2009; Proceedings, pp. 197-209, Springer, Boston.<br />

Ende, B., Gomber, P., Lutat, M. and Weber, M.C. 2010. A methodology to assess the benefits<br />

of smart order routing, Proceedings of IFIP I3E 2010 conference.<br />

Fidessa 2010. Fidessa Fragmentation Index May 2008 until April 2010,<br />

http://fragmentation.fidessa.com, accessed 04th May 2010.


Foucault, T. and Menkveld, A. 2008. Competition for Order Flow and Smart Order Routing<br />

Systems, Journal of <strong>Finance</strong>, 63, 1, pp. 119-158.<br />

Gomber, P. and Chlistalla, M. 2008. Implementing MiFID by European execution venues –<br />

Between threat and opportunity, Journal of Trading, Spring 2008, 3, 2, pp. 18-28.<br />

Gomber, P., Chlistalla, M., Gsell, M. and Pujol, G., Steenbergen, J. 2007. Umsetzung der<br />

MiFID in Deutschland - Empirische Studien zu Status Quo und Entwicklung der MiFID-<br />

Readiness der deutschen Finanzindustrie, Books on Demand.<br />

Hengelbrock, J. and Theissen, E. 2009. Fourteen at One Blow: The Market Entry of Turquoise<br />

(December 31, 2009). Available at SSRN: http://ssrn.com/abstract=1570646.<br />

Huang, R.D. and Stoll, H.R. 1996. Dealer versus auction markets: A paired comparison of<br />

execution costs on NASDAQ and the NYSE, Journal of Financial Economics, 41, 3, pp.<br />

313-357.<br />

Karsch, W. 2007. Top 100 der deutschen Kreditwirtschaft: Auf Wachstumskurs, Die Bank,<br />

2007, 8, pp. 34-37.<br />

Kundisch, D. and Holtmann, C. 2008. Competition of Retail Trading Venues – Onlinebrokerage<br />

and Security Markets in Germany. in: F. Schlottmann, D. Seese, C. Weinhardt, (Eds.),<br />

Handbook on Information Technology in <strong>Finance</strong>, pp. 171-192, Springer, Berlin.<br />

Macey, J. and O’Hara, M. 1996. The Law and Economics of Best Execution, Journal of<br />

Financial Intermediation, 6, 3, pp. 188-223.<br />

McCleskey, S. 2004. Achieving Market Integration - Best Execution, Fragmentation and the<br />

Free Flow of Capital, Butterworth-Heinemann, Elsevier.<br />

Oxera 2009. Monitoring prices, costs and volumes of trading and post-trading services, report<br />

prepared for European Commission DG Internal Market and Services, July 2009.<br />

Riordan, R., Storkenmaier, A. and Wagener, M. 2010. Fragmentation, Competition and Market<br />

Quality: A Post-MiFID Analysis (June 18, 2010), available at SSRN:<br />

http://ssrn.com/abstract=1626711.<br />

Roll, R. 1984. A simple implicit measure of the bid-ask spread in an efficient market, Journal of<br />

<strong>Finance</strong>, 39, 4, pp. 1127-1139.<br />

Stoll, H.R. 1989. Inferring the components of the bid-ask spread: theory and empirical tests,<br />

Journal of <strong>Finance</strong>, 44, 1, pp. 115-134.<br />

i The WpHG in the version of 01 November 2007, amended by the law for implementing the Markets in Financial<br />

Instruments Directive (DIR 2004/39/EG, MiFID) and the implementing directive (DIR 2006/73/EG) of the Commission (law<br />

for implementing the financial markets directive) of 16 July 2007, Federal Law Gazette I 2007, 1330 of 19 July 2007<br />

ii Ordinance Specifying Rules of Conduct and Organisation Requirements for Investment Firms (WpDVerOV) of 20 July<br />

2007, Federal Law Gazette I 2007, p. 1432 of 23 July 2007<br />

iii A segment is divided into three columns: the frequency (freq.) specifies how often an execution venue was named in total.<br />

Rank 1 and Rank 2 specify the number of times the execution venue concerned was nominated for this rank. The total of<br />

Rank 1 and Rank 2 does not necessarily match the frequency because nominations from Rank 3 and above were taken into<br />

account, but for reasons of clarity have not been included in this table. Thus, for example, Xetra for the DAX 30 values is<br />

listed six (eight) times in Rank 1 and five (five) times in Rank 2; two (one) further nomination(s) were (was) registered, but<br />

with regard to the prioritization this was counted in one of the lower ranks.<br />

iv In these cases investment firms offer order execution primarily as fixed-price business. However, if fixed-price business<br />

does not come about, the order is directed to a concrete execution venue (e.g. Xetra) which occupies Rank 2.


The Effect of Single-Stock Circuit Breakers on<br />

the Quality of Fragmented Markets<br />

Abstract: Since the May 6 th , 2010 flash crash in the U.S., appropriate<br />

measures ensuring safe, fair and reliable markets become more relevant from<br />

the perspective of investors and regulators. Circuit breakers in various forms<br />

are already implemented for individual markets to ensure price continuity and<br />

prevent potential market failure and crash scenarios. However, coordinated<br />

inter-market safeguards have hardly been adopted, but are essential in a<br />

fragmented environment to prevent situations, where main markets halt trading<br />

but stock prices continue to decline as traders migrate to satellite markets. The<br />

objective of this paper is to study empirically the impact of circuit breakers in<br />

a single market and an inter-market setup. We find a decline in market<br />

volatility after the trading halt at the home and satellite market, but at the cost<br />

of higher spreads. Moreover, the satellite market’s quality and price discovery<br />

during CBs are sorely afflicted and only restore as the other market restarts<br />

trading.<br />

Keywords: Circuit Breaker, Electronic Trading, Exchanges, Market<br />

Coordination, Market Fragmentation<br />

1. Introduction<br />

Fragmentation of investors’ order flow has been a long-time phenomenon in U.S.<br />

equity markets. Competition in European equity markets started in 2007 due to the<br />

introduction of the Markets in Financial Instruments Directive (MiFID) that enabled<br />

new venues, called multilateral trading facilities (MTF), to compete with established<br />

exchanges. Additionally, new technologies such as high frequency trading and smart<br />

order routing emerged on both sides of the Atlantic taking a significant share of the<br />

total trading volume. Against the background of these developments and the May 6 th ,<br />

2010 flash crash in the U.S., appropriate measures to ensure safe, fair and reliable<br />

electronic markets are becoming more relevant. Circuit breakers (CBs) in various<br />

forms are already implemented. However, coordinated measures between different<br />

market centers seem to become more and more relevant in the light of those recent<br />

developments. Academic research on the coordination of CBs in fragmented markets<br />

is still scarce, but would provide relevant input to market participants, exchange<br />

operators and regulators. In this paper we fill the gap by empirically investigating


CBs and their effects on liquidity, volatility and price discovery in a single market<br />

first and eventually for an inter-market case. This enables us to determine (i) how<br />

current CBs help to improve market quality for investors at the home and satellite<br />

market and (ii) if trading activity migrates from a market on halt to a satellite venue,<br />

like it is predicted in theoretical work and (iii) if this satellite venue is able to discover<br />

an effective price during the halt of the home market. Particularly, the second might<br />

pose a systemic risk to the European trading landscape if volatility shocks in one<br />

market would allow price cascades to continue in satellite trading venues, while the<br />

latter imposes risk to the integrity and reliability of fragmented markets in the case of<br />

single stock CB.<br />

The remainder of this paper is structured as follows. In the next section we will<br />

provide a review of academic papers related to our work indicating the need for more<br />

empirical research on inter-market CBs. Section 3 describes our dataset and provides<br />

some descriptive statistics for our single- and inter-market analysis of CBs while the<br />

next section presents our findings within the inter-market context. Eventually,<br />

conclusions are given in section 5.<br />

2. Related Literature<br />

Analyses of CBs in the single market case can be found widely in academic literature.<br />

Most theoretical models address the question if CBs are useful to protect market<br />

participants against extensive market volatility. [1] show how CBs may help to<br />

overcome some of the informational distortion problems caused by volume shocks in<br />

continuous trading and thereby help to improve the market's ability to absorb large<br />

volume shocks, i.e. abnormal large quantities of orders. The authors propose a<br />

temporarily switch to an alternative transactional mechanism in order to trade<br />

immediacy provided by continuous trading for information allocation in auctions.[2]<br />

prove in a model-based approach that CBs lessen the order implementation risk which<br />

happens in times of substantial volatility. [3] exhibits that CBs can be understood as a<br />

substitute for margin requirements as the down- and upside risk are limited to the<br />

corridor of the CB. In contrast to the beneficial effects described before many<br />

theoretical works on CB suggest that market activity is only delayed [4] and CB cause<br />

a magnet effect towards the threshold [5]. While the previous mentioned theoretical<br />

models on CBs deal with markets in general the following empirical studies focus on<br />

equity markets. The central question which most empirical studies tackle is how<br />

volatility is affected by CBs. Most of the studies conclude that CBs are not helping in<br />

decreasing volatility [6]. [7] does not find any support of the hypothesis that CBs help<br />

the market to calm down. [8] and likewise [9] observed a spillover effect of the<br />

volatility to the near future after a trading halt was put in place. [10] found no<br />

significant impact of CB’s on the volatility in the equity market of Bangladesh. [11]<br />

proved that consecutive halts on the Taiwan Stock Exchange dampen the volatility<br />

better than single and closing halts. Further a decreasing volatility could be observed<br />

on Korea Stock Exchange via a portfolio-based approach [12]. While academic


esearch on CBs in the case of a single market is quite extensive, research on the<br />

coordination of CBs in fragmented markets is still scarce. Focusing on inter-venue<br />

effects, [13] analyzes the utility of CBs implemented in a two-market perspective.<br />

The first market is modeled as a dominant market with a relatively high liquidity and<br />

the second market attracts only a small volume during trading hours. The authors<br />

show that - in the case of a trading halt in the dominant market - liquidity as well as<br />

price variability will shift towards the satellite market. This leads to a negation of any<br />

beneficial effects intended by the implementation of CBs on the primary market.<br />

Besides these results, the authors acknowledge that if a CB is triggered on a<br />

coordinated basis across venues, price variability decreases at the cost of declining<br />

liquidity. [14] conclude in an argumentative approach that uncoordinated CBs will<br />

more harm the market than help due to higher volatility and a rising demand in<br />

liquidity on the non-halting markets. He suggests that a better coordination across<br />

venues is strongly needed to ensure the effectiveness of such mechanisms.<br />

The first empirical study concerning the two market case is provided by [15]. In their<br />

research paper, they address market activities during NYSE halts at the NASDAQ,<br />

which is regarded as an OTC market. They find evidence that even if trading volume<br />

declines, volatility spikes significantly during these times. Due to the nature of<br />

volatility, which gives traders no arbitraging opportunity to capitalize on this<br />

situation, they conclude that a halt should not be mandatory for both trading locations.<br />

The most recent and comprehensive research on this matter was presented by [16]<br />

which is focused on delayed openings in the U.S. Their dataset consists of 2,461 halts<br />

in 1,055 stocks at the NYSE between 2002 and 2005. During these halts, trading at<br />

other venues was allowed. They find an increase of off-NYSE trading within the<br />

observation period and a significant contribution by off-NYSE trades to the price<br />

discovery process. Further they suggest that off-NYSE trades dampen the abnormal<br />

post-halt volatility and spreads. This leads them to the conclusion that continued<br />

trading may be beneficial to the market even at higher spreads. Research on CBs in<br />

two or more market case is scarce and empirical work only addresses the U.S. market<br />

system. This is especially unfavorable as empirical findings can be hardly compared<br />

to other markets due to the specialties of the U.S. trading and post-trading system.<br />

Noteworthy in the context is the Regulation NMS and its trade-trough-rule (also<br />

known as order protection rule). This rule prohibits market places to execute trades<br />

and forces them to pass through or cancel the order if the price is below/over the<br />

National Best Bid and Best Offer (NBBO). Further the implementation of the MiFID<br />

in Europe initiated a fragmentation process which makes the research in this topic<br />

necessary. Due to the lack of the European market place analysis, we are among the<br />

first who provide empirical findings for the European trading landscape within the<br />

single market and inter-market case.


3. Data Setup<br />

The absence of a mandatory inter-market coordination between MTF and regulated<br />

markets (RM) in the European market system potentially induces multiple scenarios<br />

for market quality shifts before, during and after the CB. To systematically illustrate<br />

these changes in market quality and trading behavior during a home market CB, we<br />

provide an empirical analysis of CBs effects. First, we study impact of CBs on the<br />

securities’ home market and the securities’ satellite market. Therefore we are<br />

analyzing market quality before and after the CB. This gives us the opportunity to<br />

evaluate CBs in a single market scenario, i.e. to determine its “circuit breaking”<br />

ability. Further, trading at the satellite market during the home markets halt will be<br />

assessed in order to evaluate changes in market conditions through the home market’s<br />

halt. Second, we will take a closer look at coordination effects between both markets<br />

in order to analyze trading migration and price discovery during the CB. We are<br />

among the first who provide insight to inter-market coordination effects of CBs,<br />

especially in illustrating each market’s role in the price determination process. Our<br />

sample is thereby structured as follows. We focus at instruments traded on multiple<br />

venues. In the case of a home market’s CB, trading at the satellite venue is assessed in<br />

the time interval of the CB triggered trading halt.<br />

TABLE 1: SUMMARY STATISTICS OF THE SAMPLE<br />

Number of Circuit Breakers 464<br />

Number of Instruments 27<br />

Average (median) number of CB per Instrument 17.19 (13.00)<br />

Maximum CBs per Instrument 73<br />

Minimum CBs per Instrument 4<br />

Average (median) duration in minutes 02:16 (02:16)<br />

Table 1 illustrates the aggregated descriptive statistics for our sample. Based on the<br />

German Blue Chips DAX 30 index, we analyze CBs in the year 2009 at Deutsche<br />

Boerse’s electronic order book Xetra (home market). As satellite trading venue, we<br />

choose the most relevant MTF for German Blue Chips in respect of trading volume -<br />

the London based Chi-X MTF. The 2009 scenario provided relative calm market<br />

conditions, after the 2008 financial crisis which was accompanied by tremendous<br />

market turmoil. Due to a volatility flag, we are able to identify all CBs during<br />

continuous trading over the given time period. The sample consists of the<br />

millisecond-precise start and end point of each interruption on Xetra. The total sample<br />

size consists of 522 single stock CBs in the DAX 30 instruments. Modifications<br />

within the DAX 30 index composition in 2009 and Volkswagen’s volatile price<br />

movement caused by the merger attempt by Porsche made it necessary to exclude 3<br />

instruments (58 CBs). This leads to a total sample of 464 CBs in 2009 distributed<br />

over 27 stocks. We retrieved reference data from the Thomson-Reuters Data Scope<br />

Tick History covering all Chi-X DAX 30 trades in the respective time horizon. Since


the CB data includes the start and end point of each interruption, time and date<br />

mapping will reveal trading activity on Chi-X during the home markets halt due to a<br />

CB. Based on our data, in 16% of all CBs we observe no trades on Chi-X, i.e. in 84%<br />

of all cases at least one trade occurred. Chi-X will generally not halt their electronic<br />

order book in case of a reference market’s CB. Nevertheless, Chi-X provides no<br />

information whether trading was halted during the home markets interruption [17]. In<br />

order to get consistent analysis and due to the competitive structure of the European<br />

trading market, we have to assume that Chi-X will continue trading whenever the<br />

trading halt is not triggered by a regulatory intervention.<br />

4. Results<br />

Single Market Case<br />

A trade price reaching either the last price’s dynamic threshold or the last auction<br />

price’s static threshold will trigger a CB and the trading phase will either halt or<br />

switch to auction mode. Thereby, this 2 minute interruption is meant to calm down<br />

unreasonable price movement caused by information overload. At first, we want to<br />

determine if CBs are actually capable in calming down above-average trading activity<br />

by analyzing market quality parameters before and after a CB on the home market. To<br />

do so, we look at price volatility around the CB as proposed by [11]. We measure<br />

price volatility in two ways. High variability in asset prices indicates large uncertainty<br />

about the value of the underlying asset, thus alienating an investor’s valuation and<br />

potentially resulting in incorrect investment decisions when price variability is high<br />

[18]. We obtain standard deviation of trade prices to account for the average prices’<br />

waviness around its mean. On the other hand we use a high to low ratio in order to<br />

take a closer look at the prices’ maximum deviation, i.e. the amount of mispricing.<br />

Both values outline different characteristics of volatility, so both are included in the<br />

analysis. Further, we calculate the relative spread, i.e. the relative differences between<br />

the best bid and best ask quote. This allows us to measure the risk premium market<br />

participants require for being exposed to market risk while submitting orders to the<br />

order book. [18] relates relative spread with overall market liquidity and<br />

effectiveness, therefore we also calculate the relative spread before and after the CB.<br />

All measures are calculated before the CB was triggered and after continuous trading<br />

restarted. We choose a short term 2 minutes interval, a medium term 5 minutes<br />

interval and a long term 10 minutes interval. These intervals where chosen, because<br />

the average CB in our sample interrupts trading for 02:16 minutes, we find it<br />

unreasonable to check for market quality changes in a future distance, since these<br />

changes could not be directly related to the CB. Pre- and post-CB market quality<br />

parameters are compared via Wilcoxon sign rank test for equality of the pre- and post-<br />

CB sample [19]. Table 2 depicts results for the 2 minute interval, due to space<br />

constrains the 5 and 10 minute interval could not be included, results however remain<br />

the same.


TABLE 2: MARKET QUALITY PARAMETER BEFORE AND AFTER THE CB ON<br />

XETRA<br />

2min standard deviation<br />

2min high low ratio<br />

2min relative spread (in Bps)<br />

Before After Z-Value<br />

0.064<br />

(0.035)<br />

0.010<br />

(0.007)<br />

19.33<br />

(11.77)<br />

0.058<br />

(0.035)<br />

0.010<br />

(0.006)<br />

22.06<br />

(12.65)<br />

2.811***<br />

1.362<br />

-4.730***<br />

Average (median) market quality parameters before and after the CB on Xetra for all DAX 30<br />

instruments. Z-Values for Wilcoxon sign rank test under the null that samples are drawn from<br />

the same population. Significance level are 1% (***), 5% (**) and 10% (*).<br />

The results illustrate a highly significant drop in the market volatility parameters after<br />

the CB. Besides the 2min high low ratio, we find all standard deviations as well as<br />

high low ratios on a lower level within the 2 minutes, the 5 minutes and the 10<br />

minutes interval, indicating calmer trading conditions after the CB in the short- as<br />

well as in the long-run. This comes by the cost of a higher implicit risk premium<br />

which is required for market participants after the CB, i.e. a higher spread level. We<br />

find the relative spread significantly increased in all analyzed intervals revealing an<br />

adoption of the overall market premium after the CB. First intuition for this market<br />

change one would suspect a decline in trading activity to be the driver of this shift.<br />

[20], [21], [22], and [23], among others, document a positive relation between<br />

volatility and trading volume. Therefore we look at the respective number of trades in<br />

each measurement period. By comparing the pre and post-CB trading activity for each<br />

instrument this assumption is rejected, since Wilcoxon sign rank test indicates an<br />

systematical increase in the number of trades after the CB within each observed<br />

period(results omitted due to space constrains). Calmer market volatility is achieved<br />

although trading continues on a higher activity level. Therefore, we conclude that CBs<br />

succeed to calm a home market’s hastiness by lowering market volatility after the CB.<br />

Further, this effect comes at the price of a higher spread level, which will be paid by<br />

the market participants trading after a CB.<br />

Within the current European regulatory environment alternative exchanges are<br />

allowed to continue trading even if the home market is halted due to a CB. In this case<br />

the question for trading efficiency within this period rises as inter-market coordination<br />

is not possible. We address this question by analyzing the above presented market<br />

quality parameters for the satellite market.


Again, we compare previous and posterior market quality levels using the same<br />

approach we analyzed the home market. Additionally, we measure standard deviation,<br />

high low ratio and the relative spread during the home markets interruption in order to<br />

provide insight to trading quality in this period. Therefore, we analyze CB trading in<br />

relation to the previous and subsequent CB period. As the following section will<br />

illustrate, during the CB the trading activity on Chi-X is massively reduced, so if we<br />

compare volatility on Chi-X during the CB with pre- or post-CB intervals of the same<br />

length, the results will be seriously biased as volatility will be significantly lower due<br />

to the reduced trading volume. Therefore, we recalculate the volatility measures on a<br />

per-trade manner. This way we compare values on an equal number of trades and do<br />

not suffer from a systematic bias. The before, during and after samples are tested by<br />

non-parametric Wilcoxon test for equality of medians, testing for the null that all<br />

samples are retrieved from the same population. Since every sample is tested twice,<br />

alpha levels are corrected by Bonferroni’s approach (while a given alpha value may<br />

be appropriate for each individual comparison, it is not for the set of all comparisons).<br />

In order to avoid a lot of spurious positives, the alpha value needs to be lowered to<br />

account for the number of comparisons being performed [24]. Table 3 summarizes the<br />

results of the before and after comparison as well as the between comparison<br />

approach, again 5 and 10 minute intervals where omitted due to space limitations but<br />

show identical results.<br />

TABLE 3: MARKET QUALITY PARAMETER BEFORE, DURING AND AFTER THE CB<br />

ON CHI-X<br />

Before During After Z-Value<br />

2min standard deviation<br />

0.063<br />

(0.037)<br />

0.052<br />

(0.030)<br />

4.631***<br />

2min high low ratio<br />

0.009<br />

(0.005)<br />

0.008<br />

(0.004)<br />

4.162***<br />

2min relative spread (in Bps)<br />

45.87<br />

(11.48)<br />

28.10<br />

(12.51)<br />

-3.824***<br />

standard deviation<br />

0.027<br />

(0.016)<br />

0.054<br />

(0.017)<br />

0.019<br />

(0.011)<br />

high low ratio<br />

0.012<br />

(0.002)<br />

0.011<br />

(0.003)<br />

0.005<br />

(0.002)<br />

relative spread (in Bps)<br />

50.20<br />

(13.19)<br />

133.22<br />

(50.83)<br />

37.76<br />

(14.99)<br />

Average (median) market quality parameters before, during and after the CB on Chi-X for all<br />

DAX 30 instruments Z-Values for Wilcoxon sign rank under the null that samples are drawn<br />

from the same population. Significance level are 1% (***), 5% (**) and 10% (*).


The results indicate pre- and post-CB market conditions on the satellite market to be<br />

calmer, by the meaning of a significantly reduced standard deviation as well as a<br />

lower degree of maximum price deviation. Looking at the relative spread, the home<br />

market’s findings are supported, as the satellite market also demands a higher risk<br />

premium for market risk after the CB. Again, we check the trading activity before and<br />

after the CB in order to find possible explanation in a reduced trading activity. Results<br />

indicate an increase in trading activity after the CB contradicting the before<br />

mentioned findings of related studies – again we find market condition calmer<br />

although trading activity increases. By looking at the time period during the home<br />

market’s CB, shown in the lower area of Table 4, we find prices during the<br />

interruption to be undergone by a high degree of uncertainty, compared to the period<br />

short before and short after. Standard deviation as well as the high low ratio exceeds<br />

prior and past trading levels significantly (Table 4). We also find the relative spread<br />

significantly increased. Market participants demanded a higher level of compensation<br />

for market risk as in the period the home market was active. The relative spread<br />

during the CB is on average three times as high as it is on average within the next ten<br />

minutes, and two times as high, as it was on average ten minutes before.<br />

Acknowledging, that this comparison controls for different trading volumes, we<br />

summarize a massive disturbance within the satellite market’s trading quality during<br />

the home market halt. While afterwards, with the home market proceeding with<br />

continuous trading, market quality is reassuring to a level, lower than the pre-CB<br />

period. Again this effect is accompanied by a higher spread level after the restart of<br />

the home market.<br />

TABLE 4: WILCOXON SIGN RANK TEST FOR THE EQUALITY OF MARKET<br />

QUALITY PARAMETERS BEFORE, DURING AND AFTER THE CB<br />

H0: MQPpre CB = MQPduring CB vs. H0: MQPpost CB = MQPduring CB vs.<br />

H1: MQPpre CB ≠ MQPduring CB H1: MQPpost CB ≠ MQPduring CB<br />

standard deviation -3.353*** -6.099***<br />

high low ratio -3.949*** -4.740***<br />

relative spread -13.459*** -12.773***<br />

Z-Values and significance values for Wilcoxon sign rank test for the equality of the pre-CB and<br />

during-CB market quality parameters (MQP), as well as the post-CB and during-CB sample.<br />

Bonferroni modified significance levels are 0.5% (***), 2.5% (**) and 5% (*) for each test.<br />

While previous analysis mainly focused on the independent, venue specific aspect of<br />

the CB’s impact on market quality, the following chapter will highlight inter-market<br />

coordination of trading behavior and price adjustments. In the first place, we analyze<br />

trading behavior on Chi-X during the CB. This issue is motivated by theoretical<br />

models about trading migration during CBs and the results will essentially determine<br />

the following approach to ask for price efficiency on Chi-X during the CB. Therefore,<br />

in the second step, we investigate the satellites market’s ability to uncover an efficient<br />

price in the absent of the home market.


Inter-Market Case<br />

Referring to [13], volatility shocks would allow price cascades to continue on satellite<br />

trading venues and negated any positive CB effect on the home market if traders<br />

migrate to satellite trading venues in order to continue trading. As of 2012 the<br />

European market system does not force systematic halts in case of a single venue CB<br />

halt, this scenario could occur in the Europe system. In order to control for abnormal<br />

trading behavior on Chi-X, i.e. a significant change in the number of trades, we<br />

calculated “ordinary” trading statistics within a symmetrical 30 trading day interval of<br />

the event. For each of the 30 trading days we gathered number of trades for each<br />

instrument of the corresponding CB time span. This way we can distinguish whether<br />

the trading behavior within the halt could be considered “abnormal” or “ordinary”.<br />

The length of the trading day intervals represent a short and long term view, in order<br />

to control for short term effects shortly after/before the CB and a long term “ordinary”<br />

trading behavior. Due to the extreme variability within the reference samples,<br />

parametric testing is impractical as hypothesis for normality within the data is<br />

rejected. To strengthen our implication, we perform a regression analysis. Thereby,<br />

the number of trades on Chi-X during the home market’s CB are regressed on the on<br />

the 3 day (15 day) median number of trades of the reference period. If there is actual<br />

trading migration toward the satellite market, we would expect the coefficient to be<br />

significantly larger than one. On the other side, the retreat from trading would be<br />

indicated by a coefficient significantly smaller one. We perform ordinary least square<br />

regression, suppressing the intercept and clustering for each instrument. If the null<br />

hypothesis, that the slope equals one, could not be refused, we conclude that there is<br />

no difference between the number of trades during the CBs and the reference period.<br />

In both cases, the null is rejected on a high significant level, therefore our assumption<br />

is strengthened that traders retreat from trading on the satellite market, if the home<br />

market halts (Table 5).<br />

TABLE 5: TRADING MIGRATION REGRESSION<br />

Coefficient<br />

Clustered<br />

Std. Err.<br />

T-value R 2<br />

+/- 3 day median 0.608 0.148 -2.66*** 0.3389<br />

+/- 15 day median 0.691 0.166 -1.86** 0.3242<br />

Trading migration regression to determine trading migration to Chi-X during the CB. The<br />

number of trades on Chi-X during the CBs is regressed on the 3 day (15 day) median number of<br />

trades. T-values result from a t-test under the null hypothesis of the coefficient being one.<br />

Significance level are 1% (***), 5% (**) and 10% (*).<br />

In the previous chapter we empirically show that with the home markets eruption the<br />

satellite market suffers significantly in market quality, i.e. volatility rises. While there<br />

is no trading migration within the CB interval, trading activity is actually decreasing<br />

below the expected “ordinary” level. The Question rises, if the satellite market retains


his ability to determine an efficient price during the home market’s CB and therefore<br />

still contributes to price determination, or, if the satellite market systematically fails in<br />

this ability. Therefore we look at inter-market price coordination after the CB, in<br />

order to analyze possible systematic shifts between the markets. By looking at the<br />

price coordination after the CB, we try to investigate which price, i.e. the home<br />

market’s auction price or the prices revealed on the satellite market, remains the most<br />

relevant for further trading. In order to answer this question, we start by comparing<br />

price levels on both markets in the time before and after the CB. Since a trade by<br />

trade analysis is unreasonable due to the un-harmonized trade occurrence, we<br />

calculate a venue’s per second average price and therefore rely on a per-second<br />

comparison of both markets. We do not expect both average prices to be at the same<br />

level, but by looking at this difference ratio before and after the CB, we hope to<br />

acquire possible systematic shifts, either finding the difference to be significantly low<br />

or higher the time the home market returns to continuous trading.<br />

14 Bps<br />

12 Bps<br />

10 Bps<br />

8 Bps<br />

6 Bps<br />

4 Bps<br />

2 Bps<br />

0 Bps<br />

Circuit Breaker<br />

Average Median<br />

Figure 1: Average and median differences within an eight second interval before and after the<br />

CB<br />

Figure 1 illustrates the average and median differences within an eight second interval<br />

before and after the CB. It is obvious that after the CB the difference between both<br />

venues is broader and short after approaches to the normal, i.e. pre-CB level. Table 6<br />

summarizes Wilcoxon sign rank test for the differences between each post-CB<br />

differences and the median pre-CB level. The results show that four seconds after a<br />

CB are needed for the two venues to get to the pre-CB difference level. Therefore we


assume that coordination on either one or both markets is needed after the CB in order<br />

to achieve a normal level in price difference on both markets. This interesting fact<br />

raises a new question: Which price is the most relevant for future trading?<br />

TABLE 6: TRADE PRICE DIFFERENCES BETWEEN XETRA AND CHI-X BEFORE AND<br />

AFTER THE CB<br />

Relative differences in Bps<br />

Z-Value<br />

mean (median)<br />

Pre-CB level 8.96 (1.20) -<br />

CB + 1 second period 11.80 (2.14) -5.340***<br />

CB + 2 second period 10.84 (1.69) -3.798***<br />

CB + 3 second period 9.84 (1.63) -2.154**<br />

CB + 4 second period 9.71 (1.34) -1.335<br />

CB + 5 second period 9.37 (1.33) -1.611<br />

CB + 6 second period 9.12 (1.26) -1.657<br />

Relative differences in % between the home and satellite market after the CB. Z-Value contains<br />

statistics for a Wilcoxon sign rank test with the null hypothesis that samples are obtained from<br />

the same population. Significance level are 1% (***), 5% (**) and 10% (*).<br />

We assume two possible scenarios for the post-CB price coordination. On the one<br />

hand, if only one price, either the home market’s auction price or the satellite markets<br />

last price during the CB, provides superior information, we would expect the other<br />

venues price to approach to this price level as both markets continue trading. On the<br />

other hand, if both prices contain different and relevant information about the future<br />

price level, we would expect both venues’ prices to adjust to a newer level. By taking<br />

a future reference price for each market, we regress both reference prices on the home<br />

market’s auction price as well as the satellite market’s last price during the CB. If<br />

there is one price with superior information, we would expect this price to be crucial<br />

in explaining both markets’ reference prices. Since the home markets auction price is<br />

highly correlated with the satellite market’s last price, we expect all coefficients to<br />

equal one, so a simple test for differences might not indicate structural differences.<br />

Instead, we systematically compare regression accuracy and model efficiency in order<br />

to reach a conclusion. If one price provides more explanatory power or less deviation<br />

than the other, we expect this model to be superior. Regression accuracy is measured<br />

in root-mean-square error (RMSE), i.e. the standard error of the estimate in the<br />

regression analysis, model explanatory power is measured in Akaike information<br />

criterion. The regressions take the following form:


P = β , ∗ P + ε ,<br />

P = β , ∗ P + φ , j ∈ auction, satellite!<br />

Where Pj is either the home markets auction price or the satellite market’s last price<br />

during the CB, PXetra and PChi-X are the venue specific reference prices, in our case the<br />

ten minute average price after the CB. Again we cluster for instruments.<br />

TABLE 7: POST-CB PRICE DETERMINATION REGRESSION<br />

Home Market’s<br />

price determination<br />

(P )<br />

(1)<br />

Satellite Market’s<br />

price determination<br />

(P )<br />

Home market’s auction price (P $% &') 0.205 (-138.91) 0.203 (-146.43)<br />

Satellite market’s first post-CB price (P ( )) ) 0.406 (432.99) 0.408 (436.23)<br />

Xetra and Chi-X reference prices are regressed on the home markets auction price and the last<br />

trade price on Chi-X during the CB. The table shows RMSE (AIC) for each regression.<br />

We expect both, the home markets auction price and the satellite market’s last price<br />

during the CB, to be good predictors for both future reference prices, since price<br />

deviation within ten minutes is limited. By looking at the individual information<br />

criteria and the RMSE, we interestingly find the Home market’s auction price to be of<br />

superior value for both markets future prices than the satellite markets price (Table 7).<br />

Thus, even the home market’s auction price explains the future Chi-X price more<br />

efficiently than the satellite market’s last price during the CB does. We repeat our<br />

regression by estimating the auction price’s and satellite price’s contribution together.<br />

We are aware, that multicollinearity may inflate our standard errors, but if the home<br />

market’s auction price systematically delivers additional explanatory power next to<br />

the information both prices share, the auction price should indicate significance while<br />

the last price during the CB should not. Regression is as follows:<br />

P * = β +,* ∗ P $% &' + β ,,* ∗ P ( )) + ε * , - ∈ ./012, 3ℎ- − .! (2)<br />

Pi denotes either the home or satellite markets 10 minutes average reference price,<br />

Pauction and Psatellite are the home markets auction price and the satellite market’s last<br />

price during the CB.


TABLE 8: COMBINED POST-CB PRICE DETERMINATION REGRESSION<br />

Home Market’s<br />

price determination<br />

(P )<br />

Satellite Market’s<br />

price determination<br />

(P )<br />

Home market’s auction price (P $% &') 0.913*** 0.920***<br />

Satellite market’s first post-CB price (P ( )) ) 0.087 0.080<br />

Xetra and Chi-X reference prices are regressed on the home markets auction price and the last<br />

trade price on Chi-X during the CB. Table shows coefficients and significance levels 1% (***),<br />

5% (**) and 10% (*).<br />

Table 8 supports our findings, the home markets auction price contains additional<br />

information in predicting both market’s future prices and therefore turns out to be<br />

significant. The satellite market’s price information are therefore redundant, i.e. not<br />

significant. This effect is only possible, if the satellite market’s last price during the<br />

CB experienced a systematic shift towards the home markets price level after the CB.<br />

Therefore, we conclude that the home market’s auction price is the dominant after a<br />

home markets CB.<br />

In summary, the significantly reduced trading activity during the CB compared to<br />

regular trading activity within the reference days levels the spread maker makers<br />

demand to continue quoting during the CB. Interestingly, the reduced trading activity<br />

does not calm down price volatility. We find standard deviation as well as maximum<br />

price uncertainty dramatically increased. This also impacts price determination at the<br />

satellite market. The results indicate that the price revelation process on the satellite<br />

market is seriously restricted during the CB. Although trading during the CB is not<br />

totally misguided, only with the restart of the home market and the return of the<br />

traders, the satellite market regains the ability to effectively reveal prices and<br />

approaches to the home market’s price level.<br />

5. Conclusion<br />

While the U.S. trading landscape has been fragmented since the late 1990s, the<br />

trading industry in Europe only recently experienced a sea change from regulatory<br />

and technological developments. Regulatory changes in the form of the Markets in<br />

Financial Instruments Directive (MiFID) laid the foundation for more competition<br />

between execution venues in late 2007, while trading was traditionally consolidated in<br />

national exchanges before. Since then, a multitude of Multilateral Trading Facilities<br />

MTFs emerged offering pan-European trading. It is common to exchanges and<br />

satellite trading platforms to apply some form of CB to ensure price continuity.<br />

Despite the numerous implementations of CBs in European primary markets, a<br />

coordination of CBs is completely missing. As shown in some theoretical models, this


situation could lead to price-falling-cascades if traders migrate from a market on halt<br />

to satellite venues which keep their systems open for investors’ order flow. Against<br />

this background, we analyzed trading of German blue chip stocks in two markets, i.e.<br />

Xetra and the most relevant European MTF Chi-X. We particularly investigated those<br />

periods when the Xetra trading system was on halt and Chi-X sustained its trading<br />

services.<br />

We find price volatility to decline significantly in both markets after a CB, while<br />

trading activity remains at a high level. In contrast to this, relative spreads do not<br />

revert to their pre-CB levels, but even increase after a trading halt. Therefore we<br />

conclude, that CBs do “break the circuit”, but there is no such thing as a free lunch.<br />

The price is paid by a higher relative spread. Moreover, we find trading in the satellite<br />

market to be sorely afflicted during CBs, while price volatility and relative spreads<br />

are tremendously increased on the satellite market. Further, we find no trading<br />

migration between venues in the number of trades. Instead, the number of trades<br />

significantly drops below an anticipated normal level during a stock’s trade<br />

interruption. Evidently, traders retreat from the satellite venue. In respect to previous<br />

academic work on the issue of CBs we reject the finding of the inter-venue migration<br />

model from [13] in case of non-coordinated market safeguards. A potential<br />

explanation for our empirical observation is that investors are reluctant to trade when<br />

the dominant market is absent as a liquidity pool due to a higher anticipated risk.<br />

Additionally, momentum traders and statistical arbitrage strategies depend on intermarket<br />

price differences. These types of market participants also have no incentive to<br />

continue trading, resulting in an elimination of another significant part of the original<br />

order flow. Our findings are in line with [25] which investigate liquidity providing<br />

strategies during extreme market movements and provide another explanation for<br />

investors’ market withdrawal. Implicit trading costs become very high and make<br />

trading during these periods unattractive to investors. This might encourage a<br />

negative impact on the satellite market’s price discovery process. We observe<br />

abnormally wide price differences between home and satellite market after the CB,<br />

making price coordination with the home market inevitable. By comparing satellite<br />

and home market reference prices, we find the satellite market price level to adjust<br />

towards the home market as trading activity raises after the home markets CB,<br />

indicating the home market’s auction price to be the dominant price for future trading.<br />

Our findings are particularly interesting for regulators, who might support instrumentbased<br />

market-wide trading halts to be necessary to ensure market quality in a<br />

fragmented landscape, as well as traders, adapting trading behavior in times of market<br />

stress.<br />

References<br />

1. Greenwald, B., Stein, J.C.: Transactional Risk, Market Crashes, and the Role of Circuit<br />

Breakers. The Journal of Business 64 (4), pp. 443--462 (1991)<br />

2. Kodres, L. E., O'Brien, D. P.: The Existence of Pareto-Superior Price Limits. The American<br />

Economic Review 84 (4), pp. 919--932 (1994)


3. Brennan, M.: A theory of price limits in futures markets. Journal of Financial Economics 16<br />

(2), pp. 213--233 (1986)<br />

4. Coursey, D. L., Dyl, E. A.: Price limits, trading suspensions, and the adjustment of prices to<br />

new information. Review of Futures Markets 9 (2), pp. 342--360 (1990)<br />

5. Ackert, L., Church, B., Jayaraman, N.: An experimental study of circuit breakers: The effect<br />

of mandated market closures and temporary halts on market behavior. Journal of Financial<br />

Markets 4 (2), pp. 185--208 (2001)<br />

6. Kim, Y. H., Yang, J. J.: What Makes Circuit Breakers Attractive to Financial Markets? A<br />

Survey. Financial Markets, Institutions & Instruments 13 (3), pp. 109--146 (2004)<br />

7. Chen, Y.-M.:X: Price limits and stock market volatility in Taiwan. Pacific-Basin <strong>Finance</strong><br />

Journal 1 (2), pp. 139--153 (1993)<br />

8. Kim, K. A., Rhee, S. G.: Price Limit Performance: Evidence from the Tokyo Stock<br />

Exchange. The Journal of <strong>Finance</strong> 52 (2), pp. 885--901 (1997)<br />

9. Bildik, R., Gülay, G.: Are price limits effective? Evidence from the Istanbul Stock<br />

Exchange. Journal of Financial Research 29 (3), pp. 383--403 (2006)<br />

10. Hassan, M. K., Islam, A. M., Basher, S. A.: Market Efficiency, Time-Varying Volatility and<br />

Equity Returns in Bangladesh Stock Market. Working Paper (2000)<br />

11. Kim, Y. H., Yang, J. J.: The effect of price limits on intraday volatility and information<br />

asymmetry. Pacific-Basin <strong>Finance</strong> Journal 16 (5), pp. 522--538 (2008)<br />

12. Lee, S.-B., Kim, K.-J.: The effect of price limits on stock price volatility: Empirical<br />

evidence in Korea. Journal of Business <strong>Finance</strong> & Accounting 22 (2), pp. 257--267 (2006)<br />

13. Subrahmanyam, A.: Circuit breakers and market volatility: A theoretical perspective.<br />

Journal of <strong>Finance</strong> 49 (1), pp. 237--254 (1994)<br />

14. Morris, C. S.: Coordinating Circuit Breakers in Stock and Futures markets. FRB Kansas<br />

City - Economic Review 75 (2), pp. 35--48 (1990)<br />

15. Fabozzi, F. J., Ma, C. K.: The over-the-counter market and New York Stock Exchange<br />

trading. The Financial Review 23 (4), pp. 427--437 (1988)<br />

16. Chakrabarty, B., Corwin, S. A., Panayides, M. A.: When a halt is not a halt: An analysis of<br />

off-NYSE trading during NYSE market closures. Journal of Financial Intermedation 20 (3),<br />

pp. 361--386 (2011)<br />

17. Chi-X: Europe Guidance Note 1; “ERRONEOUS TRADES”. Retrieved October 20, 2011,<br />

from http://www.chi-xeurope.com/chi-x-pdf-downloads/guidance-notes-1.-8-final-<br />

(clean).pdf (2011)<br />

18. Harris, L.: A transaction data study of weekly and intradaily patterns in stock returns.<br />

Journal of Financial Economics 16 (1), pp. 99--117 (1986)<br />

19. Wilcoxon, F.: Individual Comparisons by Ranking Methods. Biometrics Bulletin 6 (1), pp.<br />

80--83 (1945)<br />

20. French, K. R., Roll, R.: Stock Return Variances: The Arrival of Information and the<br />

Reaction of Traders. Journal of Financial Economics 17, pp. 5--26 (1986)<br />

21. Karpoff, J. M.: The Relation between Price Changes and Trading Volume: A Survey.<br />

Journal of Financial and Quantitative Analysis 22 (1), pp. 109--126 (1987)<br />

22. Schwert, W.: Why Does Stock Market Volatility Change Over Time? The Journal of<br />

<strong>Finance</strong> 44 (5), pp. 115--153 (1989)<br />

23. Stoll, H. R., Whaley, R. E.: The Dynamics of Stock Index and Stock Index Futures Returns.<br />

Journal of Financial and Quantitative Analysis 25 (4), pp. 441--468 (1990)<br />

24. Bonferroni, C. E.: Teoria statistica delle classi e calcolo delle probabilit. Pubblicazioni del R<br />

Istituto Superiore di Scienze Economiche e Commerciali di Firenze 8, pp. 3--62 (1936)<br />

25. Goldstein, M., Kavajecz, K.: Trading strategies during circuit breakers and extreme market<br />

movement. Journal of Financial Markets 7 (3), pp. 301--333 (2004)


High Frequency Trading<br />

Costs and Benefits in Securities Trading and its Necessity<br />

of Regulations<br />

DOI 10.1007/s12599-012-0205-9<br />

The Authors<br />

<strong>Prof</strong>. <strong>Dr</strong>. Christoph Lattemann<br />

School of Humanities and Social<br />

Sciences<br />

Jacobs University Bremen<br />

28759 Bremen<br />

Germany<br />

c.lattemann@jacobs-university.de<br />

<strong>Prof</strong>. <strong>Dr</strong>. Peter Loos (�)<br />

IWi at DFKI<br />

Saarland University<br />

66123 Saarbrücken<br />

Germany<br />

loos@iwi.uni-sb.de<br />

<strong>Dr</strong>. Johannes Gomolka<br />

Tempelhove Research<br />

10117 Berlin<br />

Germany<br />

<strong>Prof</strong>. <strong>Dr</strong>. Hans-Peter Burghof<br />

Arne Breuer<br />

<strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber<br />

Michael Krogmann<br />

<strong>Dr</strong>. Joachim Nagel<br />

Rainer Riess<br />

<strong>Prof</strong>. <strong>Dr</strong>. Ryan Riordan<br />

<strong>Dr</strong>. Rafael Zajonz<br />

Published online: 2012-03-06<br />

This article is also available in German<br />

in print and via http://www.<br />

wirtschaftsinformatik.de: Lattemann<br />

C, Loos P, Gomolka J, Burghof<br />

H-P, Breuer A, Gomber P, Krogmann<br />

M, Nagel J, Riess R, Riordan<br />

R, Zajonz R (2012) High Frequency<br />

Trading. Kosten und Nutzen im<br />

Wertpapierhandel und Notwendigkeit<br />

der Marktregulierung. WIRT-<br />

SCHAFTSINFORMATIK. doi: 10.1007/<br />

s11576-012-0311-9.<br />

© Gabler Verlag 2012<br />

1Introduction<br />

Recent publications reveal that high frequency<br />

trading (HFT) is responsible for<br />

10 to 70 per cent of the order volume in<br />

stock and derivatives trading (Gomber et<br />

al. 2011; Hendershott and Riordan 2011;<br />

Zhang 2010). This observation leads to<br />

a controversial debate over positive and<br />

negative implications of HFT for the liquidity<br />

and efficiency of electronic securities<br />

markets and over the costs and<br />

benefits of and needs for market regulation.<br />

Currently the European Union<br />

(EU) is considering the introduction of<br />

a financial transaction tax to curtail the<br />

harmful effects of HFT strategies. The<br />

consideration behind this market policy<br />

is based on the assumption that the<br />

very short-term oriented HFT trading<br />

strategies lead to market frictions. This<br />

current discourse shows that the arguing<br />

parties do not homogeneously define<br />

HFT. Reasons for this are the proponents’<br />

different but intertwined perspectives,<br />

which lead to new unanswered<br />

questions in numerous subjects of expertise.<br />

From a macroeconomic point of<br />

viewthequestionarisesifHFTconstrains<br />

or supports the allocation function of financial<br />

markets. Capital market research<br />

and information management research<br />

raise questions about the future form<br />

of intermediation in securities trading<br />

and the coming architecture of markets,<br />

about the HFT’s impact on liquidity and<br />

about price volatility. Financial authorities<br />

and regulators discuss whether HFT<br />

has a stabilizing or destabilizing function<br />

on financial systems and how a future<br />

regulation should be shaped.<br />

This collection of articles shall help to<br />

develop a common definition of HFT<br />

and contribute to the ongoing discussions.<br />

To that end we have collected articles<br />

from representatives of information<br />

systems, business management, the<br />

Deutsche Bundesbank and the Deutsche<br />

Boerse AG. The following scientists and<br />

practitioners participated in the discussion<br />

(in alphabetical order):<br />

BISE – DISCUSSION<br />

� <strong>Prof</strong>. <strong>Dr</strong>. Hans-Peter Burghof and Arne<br />

Breuer, Chair of Business Economics,<br />

especially Banking and Financial Services,<br />

University of Hohenheim, Germany.<br />

� <strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber, Chair of Business<br />

Economics, especially e-<strong>Finance</strong>,<br />

Johann <strong>Wolfgang</strong> Goethe-University<br />

of Frankfurt, Germany.<br />

� <strong>Dr</strong>. Joachim Nagel, Member of the<br />

Board of Directors, and <strong>Dr</strong>. Rafael Zajonz,<br />

Central Market Analysis, Portfolio,<br />

Deutsche Bundesbank, Frankfurt,<br />

Germany.<br />

� Rainer Riess, Managing Director of the<br />

Frankfurter Wertpapierbörse (FWB),<br />

and Michael Krogmann, Executive<br />

Vice President of Xetra Market Development<br />

of Deutsche Börse AG, Frankfurt,<br />

Germany.<br />

� <strong>Prof</strong>. <strong>Dr</strong>. Ryan Riordan, Institute for<br />

Information Systems and Management,<br />

Karlsruhe Institute of Technology<br />

(KIT), Karlsruhe, Germany.<br />

HFT is a part of algorithmic trading. Gomolka<br />

(2011) defines algorithmic trading<br />

as the processing and/or execution<br />

of trading strategies by the means<br />

of intelligent electronic solution routines<br />

(known as algorithms). Thus algorithmic<br />

trading encompasses computersupported<br />

trading as well as computergenerated<br />

sell-side and buy-side market<br />

transactions. Algorithmic trading strategies<br />

can be both short-term and longterm<br />

oriented.<br />

In general, HFT is defined as real-time<br />

computer-generated decision making in<br />

financial trading, without human interference<br />

and based on automatized order<br />

generation and order management. HFT<br />

encompasses short-term trading strategies,<br />

which – in extreme cases – operate<br />

in the range of microseconds using minimal<br />

price differences. HFT thus results in<br />

minimal profit margins per transactions<br />

and exhibits very short holding periods of<br />

securities positions.<br />

However, HFT definitions vary and<br />

various properties of HFT are not<br />

consistently discussed in the literature.<br />

Business & Information Systems Engineering 2|2012 93


BISE – DISCUSSION<br />

Aldridge’s (2009) definition of HFT holding<br />

periods range from milliseconds to<br />

one day. Durbin (2010) on the other<br />

hand describes HFT as trading strategies,<br />

which covers seconds or milliseconds<br />

only. According to Brogaard (2010),<br />

HFT is extremely short-term buying or<br />

selling with the intention to profit from<br />

minimal price fluctuations.<br />

Further characteristics are often mentioned<br />

but are not always included in<br />

HFT definitions, such as the exclusive usage<br />

by professional/institutional investors<br />

in proprietary trading, real-time data<br />

processing and direct market access (Dacorogna<br />

et al. 2001). Another controversial<br />

issue is the avoidance of overnight<br />

risk (Aldridge 2009). Other definitions<br />

underline the role of HFT as financial<br />

intermediary (Hendershott and Riordan<br />

2011) or try to find differences among<br />

the implemented trading strategies (Ye<br />

2011).<br />

On the basis of the broad HFT definition<br />

given before the authors in this article<br />

will shed light on the following questions:<br />

(1) How does HFT influence the<br />

liquidity and efficiency of electronic securities<br />

markets? (2) What are the costs and<br />

benefits of, and what are the needs for a<br />

HFT regulation?<br />

Peter Gomber analyzes HFT from a<br />

market microstructure perspective, and<br />

finds HFT to be a central element of the<br />

value creation chain in securities trading.<br />

As part of the value creation chain,<br />

HFT contributes to increased efficiency<br />

and reduced explicit and implicit transaction<br />

costs. In his eyes, regulation of HFT<br />

could lead to dramatic changes in market<br />

behavior, while an inappropriate regulation<br />

might even be counterproductive for<br />

market quality. Gomber sees the problems<br />

for profound research on HFT in the<br />

lack of data available for empirical studies.<br />

Again this leads to adverse effects in<br />

discussions of the topic in the public, in<br />

the media, and with regulators.<br />

Ryan Riordan also looks at HFT from<br />

the perspective of market microstructure<br />

and interprets HFT as one form of technological<br />

financial intermediation which<br />

contributes to the efficiency of operations<br />

in exchange trading. In his eyes,<br />

HFT plays an important role in the process<br />

of price formation and influences the<br />

size of transaction costs in securities trading.<br />

According to him, one cannot yet<br />

say whether HFT will have a positive or<br />

a negative impact on the capital markets.<br />

However, he sees major advantages in a<br />

highly technologized market. It is no alternative<br />

for him to turn back the wheels<br />

and return to a backward oriented, artificially<br />

slowed, regulated trading, which is<br />

based on human intermediation.<br />

Rainer Riess and Michael Krogmann<br />

describe HFT as the highest evolutionary<br />

level of securities trading. In their<br />

opinion HFT leads to faster processing<br />

of information, to an increase in liquidity,<br />

and thus added values for the<br />

overall economy. The authors describe<br />

how HFT is currently technically realized<br />

and integrated into trading operations<br />

at the exchange, and deduct their<br />

arguments accordingly. From the point of<br />

view of Deutsche Börse, HFT is mainly<br />

used by institutional investors in proprietary<br />

trading and focuses on highly<br />

liquid stocks. The authors correlate the<br />

rise of HFT with a continuous improvement<br />

of the electronic trading system XE-<br />

TRA, which – from the point of view<br />

of Deutsche Börse – benefits all market<br />

participants in the same way. In the eyes<br />

of Riess and Krogmann, a future regulation<br />

of HFT should primarily focus on<br />

equal chances of competition in the EUarea,<br />

in order to create “a level playing<br />

field“. From the point of view of Deutsche<br />

Börse, it is necessary not only to implement<br />

security mechanisms on the side of<br />

exchanges but also with HFT-firms.<br />

Arne Breuer and Hans-Peter Burghof<br />

also recognize that, due to HFT, information<br />

can be processed more perfectly<br />

and faster than ever before. They look at<br />

the topic from the perspective of financial<br />

economics. This point of view leads<br />

them to believe that more and faster information<br />

does not necessarily lead to<br />

a correct determination of the intrinsic<br />

value of financial instruments. Rather<br />

HFT processes short-term information,<br />

which primarily is made of short-term<br />

volume and short-term time series data,<br />

and thus does not contribute to the evaluation<br />

of the intrinsic values. The authors<br />

vote for a stricter regulation of HFT.<br />

However, before this can be done, more<br />

analyses should be conducted. For this,<br />

more data are necessary.<br />

Finally, Joachim Nagel und Rafael Zajonz<br />

argue from the perspective of regulators.<br />

A blanket judgment on HFT is<br />

from the regulators’ point of view neither<br />

adequate nor would it lead to improvements<br />

of the regulatory framework<br />

regarding transparency, stability, and efficiency.<br />

The impact of HFT on the efficiency<br />

of securities trading is – due to the<br />

absence of a scientific discussion – still<br />

unclear for the regulators. The possibility<br />

to destabilize the market due to HFT in<br />

volatile market situations is regarded as<br />

critical but should be looked into in detail.<br />

From the point of view of the authors<br />

“market friendly” strategies exist, a fact<br />

which can be judged positively. But there<br />

are also ”unfriendly strategies“, which –<br />

from their perspective – can be categorized<br />

as potentially harmful. In the center<br />

of their article, the authors formulate<br />

the wish that this complex topic may<br />

be discussed more intensely by the scientific<br />

community in the future, in order<br />

to better understand which fundamental,<br />

regulatory measures should be applied to<br />

HFT.<br />

If you would like to comment<br />

on this topic or another article of<br />

the journal Business & Information<br />

Systems Engineering, please send<br />

your contribution (max. 2 pages) to<br />

the editor-in-chief, <strong>Prof</strong>. Hans Ulrich<br />

Buhl, University of Augsburg,<br />

Hans-Ulrich.Buhl@wiwi.uni-augsburg.de.<br />

<strong>Prof</strong>. <strong>Dr</strong>. Christoph Lattemann<br />

School of Humanities and Social<br />

Sciences<br />

Jacobs University Bremen<br />

<strong>Prof</strong>. <strong>Dr</strong>. Peter Loos<br />

IWi at DFKI<br />

Saarland University<br />

<strong>Dr</strong>. Johannes Gomolka<br />

Tempelhove Research<br />

2 High Frequency Trading<br />

Regulation Required at a<br />

Reasonable Level<br />

It is uncommon for a specific subject<br />

in the field of securities trading and ITinnovation<br />

to draw as much public attention<br />

as high frequency trading (HFT) has<br />

been doing in recent months. Merely a<br />

special field for a small group of experts<br />

prior to 2010, it is now a frequent part<br />

of the general news coverage. Against the<br />

background of the recent debt crisis, the<br />

current volatility and market turmoil as<br />

well as the “US Flash Crash” on May 6,<br />

2010 lead to this extreme attention. Several<br />

parties attempt to exert pressure on<br />

politics and regulation by making HFT<br />

responsible for that crisis and the high<br />

market volatility. In reaction to the aforementioned<br />

incidents and to the subsequent<br />

public discussions, the regulatory<br />

94 Business & Information Systems Engineering 2|2012


authorities of international financial centers<br />

have debated the adoption of various<br />

regulatory measures and now propose<br />

regulatory procedures, which currently<br />

substantiate especially in Europe<br />

and will presumably be approved in 2012<br />

in the context of the revision of the Markets<br />

in Financial Instruments Directive<br />

(MiFID).<br />

Basically, the trading strategies based<br />

on HFT can be subdivided into active<br />

and passive ones. Whereas passive<br />

strategies provide other market participants<br />

with trading opportunities in terms<br />

of quotes and limit orders (e.g. electronic<br />

market making), active strategies<br />

primarily attempt to exploit imbalances<br />

of asset prices in fragmented markets<br />

(e.g. primary market and multilateral<br />

trading facilities), discrepancies in<br />

valuation between different asset classes<br />

(e.g. between derivatives and their underlyings)<br />

or deviances of current asset<br />

valuations compared to historical<br />

correlations (e.g. in the so-called pairs<br />

trading) immediately after the emergence<br />

of these trading/arbitrage opportunities.<br />

The emerging academic literature,<br />

which analyzes the effects of HFT based<br />

strategies on market quality, shows<br />

mostly positive impact (for a systematic<br />

outline of academic research concerning<br />

HFT see Gomber et al. 2011).<br />

Regarding price discovery, liquidity and<br />

volatility, most studies discover positive<br />

effects of HFT. Only a few publications<br />

indicate that HFT can increase the adverse<br />

selection problem under specific<br />

circumstances, and in the case of the “US<br />

Flash Crash” another survey (Kirilenko<br />

et al. 2011) reveals that HFT can increase<br />

volatility.<br />

The growing market efficiency and a<br />

reduction of explicit and implicit transaction<br />

costs triggered by HFT is an obvious<br />

issue particularly for those market<br />

participants who used to capitalize on<br />

intermediary services and broad bid/ask<br />

spreads in a formerly less efficient and<br />

less liquid trading environment. In contrast<br />

to off-exchange trading via internalization<br />

and so-called dark pools, i.e.<br />

non-transparent execution venues, HFT<br />

market-making strategies on lit markets<br />

face relevant adverse selection costs as<br />

they provide liquidity on the market<br />

without knowing their counterparties.<br />

Within their internalization systems and<br />

dark pools in the OTC field, banks and<br />

brokers are aware of their counterparties’<br />

identities and can benefit from this information.<br />

Contrary to this, HFTs in lit markets<br />

are not aware of the toxicity of their<br />

counterparts and are – analogous to market<br />

makers – exposed to the problem of<br />

adverse selection.<br />

Inappropriate regulation of HFT based<br />

strategies or an impact on HFT business<br />

models due to excessive burdens might<br />

turn out to be counterproductive and<br />

lead to unforeseeable consequences for<br />

the quality of markets. However, abusive<br />

strategies have to be combated effectively<br />

by the regulators. Particularly the analysis<br />

of the “US Flash Crash” with its discussed<br />

solution approaches can hardly<br />

be transferred to the European situation,<br />

since the issues related to the “US Flash<br />

Crash” primarily result from the US market<br />

structure. In Europe, where a more<br />

flexible best execution regime is implemented<br />

and a share-by-share volatility<br />

safeguard regime has been in place for<br />

two decades, no market quality problems<br />

related to HFT have been documented so<br />

far. Therefore, a European approach to<br />

the subject matter is required, and Europe<br />

should be cautious about addressing<br />

and fixing a problem that exists in a<br />

different market structure and thus creating<br />

risks for market efficiency and market<br />

quality.<br />

Any regulatory interventions in Europe<br />

should try to preserve the benefits of HFT<br />

while mitigating the risks as far as possible<br />

by assuring that (i) HFT firms are<br />

able to provide documentation on their<br />

algorithms upon authorities’ request and<br />

to conduct back-testing, (ii) markets are<br />

capable of handling peaks in trading activity<br />

and apply safeguards to react to<br />

technical issues of their members’ algorithms,<br />

(iii) a diversity of trading strategies<br />

prevails to prevent systemic risks,<br />

(iv) co-location and proximity services<br />

are implemented on a level playing field,<br />

(v) regulators have a complete overview<br />

of the possible systemic risks which could<br />

be triggered by HFT, and have employees<br />

who have the knowledge and the tools<br />

to assess the impact of the trading algorithms<br />

on market quality and the associated<br />

risks. Furthermore, it is crucial<br />

that market places in a fragmented environment<br />

develop coordinated safeguards<br />

und circuit breakers, which mirror the<br />

HFT reality and enable all market participants<br />

to react adequately even in market<br />

stress.<br />

Regulatory proposals demanding continuous<br />

liquidity provision by HFT in the<br />

sense of market marking obligations or<br />

BISE – DISCUSSION<br />

minimum quote lifetimes miss the mark<br />

and are not suitable to improve market<br />

stability or market integrity. They rather<br />

contribute to a decrease in market quality<br />

and higher transaction costs.<br />

At first sight, demanding obligations<br />

for HFTs to provide quotes seems an appropriate<br />

measure to tackle the problem<br />

of a sudden liquidity withdrawal. However,<br />

it is highly doubtful whether any<br />

rule can force market makers to buy in<br />

the face of overwhelming selling pressure.<br />

In such a situation they might rather<br />

take the risk of being fined for not fulfilling<br />

their obligations. Many HFT strategies<br />

are characterized by rapid closing<br />

of built-up positions to minimize risk.<br />

Hence, an obligation to provide liquidity<br />

and thereby risk capital is in sharp contrast<br />

to many HFT business models. Due<br />

to the significant regulatory costs those<br />

obligations would potentially lead to a<br />

retreat from the market and thus to a<br />

notable loss of liquidity.<br />

Also a minimum order lifetime, which<br />

at first glance appears to be useful to<br />

avoid fast order submissions and immediate<br />

cancellations, would lead to a<br />

significant change in market behavior.<br />

Market participants are then no longer<br />

able to react quickly and adequately to<br />

market-exogenous information (e.g. adhoc<br />

news) and the necessity to keep an<br />

order in the order book presents a free<br />

option for other market participants. Besides,<br />

the existence of minimum order<br />

lifetimes would lead to an implementation<br />

of trading strategies capitalizing on<br />

the“lockin”oforders.HFTwouldanticipate<br />

the accompanied risks and costs<br />

and attempt to compensate these costs<br />

with higher spreads, which again would<br />

have negative effects on market quality.<br />

In this debate it should not be neglected<br />

that speed is the key tool for HFTs’ risk<br />

management.<br />

HFT is an important factor in markets<br />

that are driven by sophisticated technology<br />

on all layers of the trading value<br />

chain. However, discussions on this topic<br />

often lack sufficient and precise information.<br />

A remarkable gap between the results<br />

of academic research on HFT and<br />

its perceived impact on markets in public,<br />

media and regulatory discussions (European<br />

Commission 2010) can be observed.<br />

Here, the provision of granular and reliable<br />

data by the industry can assist empirical<br />

research at the interface of finance<br />

and IS to provide important contributions<br />

to a reasonable regulation of HFT.<br />

Business & Information Systems Engineering 2|2012 95


BISE – DISCUSSION<br />

This regulation should eventually minimize<br />

the inherent risks of the technology<br />

in question without hindering the<br />

indisputably existing positive effects for<br />

market quality.<br />

<strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber<br />

University of Frankfurt<br />

E-<strong>Finance</strong> <strong>Lab</strong><br />

3 High Frequency Trading (HFT) –<br />

ANewIntermediary<br />

Financial markets require intermediaries<br />

to provide liquidity and immediacy for<br />

other participants. These intermediaries,<br />

often called market makers or specialists,<br />

were often afforded special status<br />

and located on the trading floor, or close<br />

to the trading mechanism of exchanges.<br />

The automation of financial markets has<br />

increased their trading capacity and intermediaries<br />

have expanded their use of<br />

technology. This has resulted in a reduced<br />

role for traditional human market makersandledtotheriseofanewintermediary,<br />

referred to as high frequency traders<br />

(HFTs).<br />

This development has been made possible<br />

by the technological innovations in<br />

recent years. HFT strategies usually make<br />

use of the high speed technologies to<br />

build up and unwind positions within<br />

milli- and microseconds. Prerequisites<br />

for this development were the reduction<br />

of system latency and the increase<br />

of computing power and data processing<br />

capabilities of computers. Next to the<br />

large investments in HFT, exchanges have<br />

also invested large amounts of money in<br />

their IT infrastructure. For example, the<br />

costs of a high-speed connection between<br />

Chicago and New York are estimated<br />

around $200,000 per mile (Forbes 2010).<br />

The question remains whether these investments<br />

are justified with regard to the<br />

increase of overall market quality and<br />

welfare that results from higher HFT<br />

activity on the market.<br />

Like traditional intermediaries HFTs<br />

hold little inventory, have short holding<br />

periods, and trade often. Unlike traditional<br />

intermediaries, however, HFTs<br />

are not granted preferential access to the<br />

market not available to others and they<br />

employ advanced and innovative technology<br />

to intermediate trading. Withoutsuchprivileges,thereisnoclearbasis<br />

for imposing the traditional obligations<br />

of market makers on HFT. The substantial,<br />

largely negative media coverage<br />

of HFT and the so called “flash crash”<br />

on May 6, 2010 raise significant interest<br />

and concerns about the role HFT play in<br />

the stability and price efficiency of financial<br />

markets. The predominantly negative<br />

coverage seems mostly unfounded.<br />

Overall, HFTs’ impact is similar to<br />

that of other intermediaries and speculators.<br />

Speculators can improve price efficiency<br />

by obtaining more information<br />

on prices and by trading against pricing<br />

errors. Manipulative strategies and<br />

predatory trading could decrease price<br />

efficiency. Reducing pricing errors improves<br />

the efficiency of prices. HFTs’ informational<br />

advantage, which is driven<br />

by the technology they employ, is shortterm.<br />

It is unclear whether or not this<br />

short-term information and intraday reductions<br />

of pricing errors facilitate better<br />

financial decisions and resource allocations<br />

by firms and investors. If the shortterm<br />

information – that HFTs price in –<br />

would not otherwise become public microseconds<br />

later, HFT clearly plays an important<br />

role (Hendershott and Riordan<br />

2011). It would be an important positive<br />

role of smaller pricing errors if these corresponded<br />

to lower implicit transaction<br />

costs by long-term investors.<br />

One important point left unaddressed<br />

thus far is whether or not HFTs engage in<br />

manipulative or predatory trading. Their<br />

use of technology may allow HFTs to manipulate<br />

prices at speeds that are undetectable<br />

by slower traders. A manipulative<br />

strategy might be the ignition of a<br />

price movement in one direction only in<br />

order to trade on the opposite side of the<br />

market as proposed by the SEC (2010)<br />

and therefore cause significant pricing<br />

errors. As is frequently done, one can<br />

argue whether the underlying problem<br />

of possible manipulation lies with the<br />

manipulator or the market participant<br />

who is manipulated. In the SEC example,<br />

the passive manipulation could not<br />

succeediftherewerenopricematching.<br />

The manipulation stories could be tested<br />

with more detailed data identifying each<br />

market participant’s orders, trading, and<br />

positions in all markets.<br />

Despitethestrongevidenceofthepositive<br />

role of HFT for the efficiency of<br />

price determination and trading costs<br />

(Hendershott et al. 2011; Brogaard 2010;<br />

Zhang and Riordan 2011), regulators and<br />

the media are certain that they must be<br />

regulated.Itis,however,unclearandalso<br />

debatable how we should regulate HFT.<br />

Assuming that some, or most, of their<br />

activities contribute positively to liquidity<br />

and price efficiency, which parts of<br />

their trading should we regulate? There<br />

are controversially discussed suggestions<br />

to restrict HFTs’ mostly passive trading or<br />

to enforce a minimum order life on limit<br />

orders. Restricting HFTs’ ability to trade<br />

actively necessarily impedes their ability<br />

to manage the risks associated with intermediation.<br />

This may lead to less intermediation<br />

and lower liquidity. Imposing<br />

minimum order lives on limit orders<br />

may also negatively impact HFTs’<br />

ability to manage trading risks during<br />

volatile market periods that existed before<br />

HFT dominated the equity market.<br />

Finally, the discussions of US and European<br />

regulation should take into account<br />

specific differences of both markets. Despite<br />

the high market fragmentation, the<br />

European market has maintained a comparably<br />

high degree of efficiency. This is<br />

also due to the help of HFTs. They make<br />

use of arbitrage strategies to dissolve existing<br />

price deviations within seconds<br />

which results in an interconnectedness of<br />

European markets.<br />

A final point is a more general one<br />

on technology investments. HFTs must<br />

makealargeandlong-terminvestmentin<br />

technology, both hardware and software.<br />

This investment in technology seems to<br />

have to paid-off both for HFTs and the<br />

equity markets. If regulation were to<br />

negatively impact the returns on investments<br />

in HFT technologies by reducing<br />

the profitability of intermediation, fewer<br />

firms will be willing to invest in these<br />

technologies. This may lead to a situation<br />

in which one or two highly specialized<br />

firms dominate intermediation,<br />

which ultimately leads to less competition,<br />

lower liquidity and reduced priceefficiency.<br />

Competition, ease of market<br />

entry and the use of specialized and innovative<br />

technology seem to be the best<br />

guarantors of market stability.<br />

It is hard to imagine a situation in<br />

which HFTs are able to artificially manipulate<br />

prices for longer periods of<br />

time given the intense competition other<br />

HFTs. HFTs are one type of intermediary.<br />

When thinking about the role HFT plays<br />

in markets it is natural to try to compare<br />

the new market structure to the previous<br />

market structure. Some primary differences<br />

are that there is free entry into<br />

HFT, HFTs do not have a designated role<br />

with special privileges, and HFTs do not<br />

have special obligations. When considering<br />

the optimal industrial organization of<br />

the intermediation sector, which includes<br />

96 Business & Information Systems Engineering 2|2012


egulation, market structure, technology<br />

and incumbency, HFT more closely resembles<br />

a highly competitive environment<br />

than traditional market structures.<br />

A central question is whether there were<br />

benefits of the more highly regulated and<br />

less technology intensive intermediation<br />

sector which outweigh the costs of lower<br />

innovation and higher entry costs typically<br />

associated with regulation. The answer<br />

to this question seems thus far to be<br />

a resounding “no”.<br />

<strong>Prof</strong>. <strong>Dr</strong>. Ryan Riordan<br />

Karlsruhe Institute of Technology<br />

4 High Frequency Trading – An<br />

Exchange Operator’s Perspective<br />

4.1 High Frequency Trading – Myth and<br />

Reality<br />

On 2010-09-30, the U.S. Securities and<br />

Exchange Commission (SEC) and the<br />

Commodity Futures Trading Commission<br />

(CFTC) (2010) issuedajointreport<br />

showing that the so-called “flash crash”,<br />

a sequence of events which made prices<br />

plunge throughout the US stock market,<br />

was caused by an incorrectly programmed<br />

trading algorithm of a traditional<br />

investment company which did<br />

not use high frequency trading (HFT).<br />

Nevertheless, HFT has gained massive<br />

public attention ever since. The news media,<br />

as well scientists and regulatory authorities,<br />

are busy discussing and analyzing<br />

the effect of HFT on the global capital<br />

markets. While the public perception<br />

of HFT is largely critical – and driven by<br />

headlines demanding a HFT ban or, at<br />

least, strict regulation – scientific analysis<br />

comes to rather different conclusions (see<br />

Gomber’s discussion above). According<br />

to Brogaard’s (2010) study of HFT, blaming<br />

HFT for the US flash crash is not<br />

the only popular fallacy regarding the<br />

role of HFT in securities trading. Brogaard’s<br />

analysis of NASDAQ data showed<br />

thatfor65%ofthetimeHFTaccounted<br />

for the best bid and ask quotes. Also,<br />

Broogardfoundnoevidencesuggesting<br />

that HFT firms systematically engage in<br />

market abuse, e.g. by illegally taking advantage<br />

of information about client orders,<br />

the so-called “front running”. Since<br />

HFT firms are proprietary traders, they<br />

do not have any clients. Generally, scientific<br />

analysis did not find a correlation<br />

between HFT and market abuse.<br />

The Netherlands Authority for the Financial<br />

Markets (AFM 2010) considersHFT<br />

as a legitimate trading method which is<br />

not market abusive under normal circumstances.<br />

According to Gomber, academic<br />

papers mostly could not find evidence<br />

for negative effects of HFT on<br />

market quality. Moreover, the Germanybased<br />

Karlsruhe Institute of Technology<br />

(KIT) concluded their study based on<br />

analysis of NASDAQ data with the finding<br />

that HFT even worked as a buffer<br />

against plunging stock prices during the<br />

crisis years 2008 and 2009 (Zhang and<br />

Riordan 2011).<br />

4.2 Insights of an Exchange Operator<br />

We live in a technology-driven society,<br />

continuously striving to further improve<br />

and advance the achievement potential<br />

of our economy as well as of nearly<br />

every aspect in our everyday life: can<br />

anyone imagine a commercial flight today<br />

without the aid of an autopilot, or<br />

modern microsurgery without robotics?<br />

These advancements are by no means<br />

ends in themselves but serve a greater<br />

good. Just the same goes for the ever<br />

increasing speed in securities trading –<br />

a development which leads to continuously<br />

improving general market quality<br />

and also to more efficient risk management<br />

for every market participant. The<br />

faster the market data transmission, the<br />

faster investors are able to adapt to ongoing<br />

market developments. This does not<br />

only have a very positive effect on the<br />

safety in securities trading but also on<br />

transaction cost: faster trading leads to<br />

tighter spreads and, therefore, to higher<br />

liquidity. The implicit transaction costs<br />

of every securities trade are determined<br />

mainly by liquidity and account for up<br />

to 80 percent of the overall transaction<br />

costs, while the explicit transaction costs<br />

– commissions, fees, taxes – are of minor<br />

significance. With this in mind, Deutsche<br />

Börse started long before the advent of<br />

HFT to improve the trading infrastructure<br />

of its electronic trading platform Xetra,<br />

especially in view of ever decreasing<br />

systemic latency. At the same time,<br />

Deutsche Börse further developed the security<br />

mechanisms and technologies respectively<br />

adapted them to the increasing<br />

demands of a more and more sophisticated<br />

and faster trading system, one of<br />

them being the very effective instrument<br />

of the volatility interruption, introduced<br />

in 1999. This security mechanism is used<br />

in extremely volatile market phases and<br />

BISE – DISCUSSION<br />

leads to higher price stability: whenever<br />

an indicative price is outside the price<br />

range – which is pre-defined for every security<br />

traded on Xetra – a volatility interruption<br />

will be initiated around the<br />

reference price.<br />

While continuously advancing the<br />

technical infrastructure, Deutsche Börse<br />

expanded its offer of individually selectable<br />

bandwidths for market participants<br />

connected to Xetra from<br />

512 Kbit/sec up to 2 Mbit/sec for their<br />

Values API interfaces. In 2008, for Xetra<br />

members requiring even faster market<br />

data transmission and more order<br />

book depth, an additional interface with<br />

a bandwidth of 1 Gbit/sec was implemented,<br />

called Enhanced Broadcast Solution<br />

respectively Enhanced Transaction<br />

Solution. Today, bandwidths of up to 10<br />

GBit/sec are available. With the introduction<br />

of the so-called “non-persistent”<br />

orders in 2009, Deutsche Börse further<br />

enabled Xetra members to optimize their<br />

response times to price changes thanks<br />

to even faster data processing. “Nonpersistent”<br />

orders are not saved in exchange<br />

systems and are thus designed not<br />

be executed after volatility interruptions.<br />

In late 2011 Deutsche Börse complemented<br />

its connectivity portfolio with<br />

the FIX (Financial Information Exchange)<br />

gateway. Market participants usingthisprotocolnowareabletoconnect<br />

to Xetra far more easily.<br />

However, there was one latency factor<br />

left that even the most sophisticated technology<br />

could not overcome: the propagation<br />

delay due to physical distance. For<br />

every 100 km which a market participant’s<br />

trading engine and the trading system<br />

of Xetra are physically apart, transaction<br />

latency increases by 1 msec approximately.<br />

This could mean a true competitive<br />

disadvantage for market participants<br />

relying on ultra low latency. Deutsche<br />

Börse addressed this growing market demand<br />

by introducing its proximity services<br />

in 2006. By placing the trading engine<br />

of distant Xetra members not only<br />

virtually but physically close to the exchange<br />

back end – a process called colocation<br />

– the travel time of the market<br />

data could be drastically reduced. Today,<br />

141 Xetra members take advantage<br />

of Deutsche Börse’s co-location offer.<br />

Thanks to a continuously perfected<br />

trading infrastructure and the introduction<br />

of proximity services, Deutsche<br />

Börse has not only remained competitive<br />

on an international level but has also prepared<br />

Xetra optimally for the needs of<br />

Business & Information Systems Engineering 2|2012 97


BISE – DISCUSSION<br />

HFT firms. Over the last few years, systemic<br />

latency on Xetra has been further<br />

reduced notwithstanding a dramatic increase<br />

of technical transactions – an advantage<br />

to all market participants alike:<br />

a fair, equal access to Xetra and the preand<br />

post-trade transparency characteristic<br />

of a regulated exchange make sure<br />

that every investor enjoys all advantages<br />

Deutsche Börse’s trading platform has to<br />

offer.<br />

While being a minority, HFT firms<br />

nevertheless play an important role in<br />

improving the order book quality on Xetra,<br />

e.g. by bundling the very heterogenic<br />

order flow. There are three organized<br />

forms of HFT on Xetra: the proprietary<br />

trading of investment firms, hedge<br />

funds, and proprietary trading companies.<br />

Two types of trading prevail: first of<br />

all, the so-called electronic liquidity provision.<br />

In this case, HFT firms act as voluntary<br />

market makers, adding liquidity<br />

to a multitude of securities. The second<br />

type of HFT on Xetra is called statistical<br />

arbitrage which leads to improved price<br />

discovery. Both types of HFT account<br />

for tighter spreads and, ultimately, improved<br />

market efficiency on Xetra. So far,<br />

Deutsche Börse could find no evidence of<br />

HFT having lead to destabilizing markets<br />

during periods of market turmoil, e.g. by<br />

strengthening trends. During the highly<br />

volatile market phase in August 2011, the<br />

trading volume on Xetra increased temporarily<br />

to 107 million transactions on<br />

one single day. Despite up to 30 volatility<br />

interruptions, the average transaction<br />

processing took only 0.4 msec longer<br />

than usual. System availability was guaranteed<br />

at all times, Xetra members did<br />

not have to face any restrictions, let alone<br />

system failure. Deutsche Börse’s market<br />

security mechanisms made sure that all<br />

trading activities could be executed properly<br />

and continuously while price stability<br />

was guaranteed even during market<br />

turmoil.<br />

Thus, Deutsche Börse succeeded in advancing<br />

the Xetra infrastructure in terms<br />

of continuously decreasing systemic latency<br />

and, at the same time, met the<br />

permanently increasing needs regarding<br />

safety and speed of its trading system<br />

even before the term HFT came up.<br />

4.3 Regulatory Recommendations<br />

Within a national economy it is the explicit<br />

function of a securities exchange to<br />

facilitate the most efficient employment<br />

of capital, ensuring best possible corporate<br />

financing and re-financing. HFT, as<br />

it is today, supports faster processing of<br />

economically relevant data and leads to<br />

higher liquidity in the trading of company<br />

shares. Thanks to a stable, highperformance<br />

trading system, Deutsche<br />

Börse was able to integrate HFT successfully<br />

and to use the positive effects of<br />

HFT to improve overall market quality.<br />

This would not have been possible without<br />

Deutsche Börse’s principle of equal<br />

access and a fair set of rules applying to<br />

every market participant trading on Xetraalike.Fromaregulatoryperspective<br />

– and keeping MiFID’s ultimate goal of<br />

creating an EU-wide “level playing field”<br />

in mind – comprehensive rules regarding<br />

HFT definitely would make sense.<br />

Therefore, Deutsche Börse supports all<br />

measures to enhance transparency, e.g.<br />

the complete registration of all market<br />

participants and a full recording of all<br />

their trading activities – traditional trading<br />

and HFT alike. The Deutsche Börse<br />

(2011) has come to the conclusion that<br />

regulatory intervention in HFT must not<br />

hurt the proven positive effect on market<br />

quality HFT has to offer. In particular,<br />

the variety of HFT strategies should<br />

be preserved, as systemic risk should be<br />

prevented. To achieve these goals, HFT<br />

firms themselves may have to implement<br />

security mechanisms – just as exchange<br />

operators as Deutsche Börse already have.<br />

Whichever regulatory rules may be implemented<br />

in the end, the regulators will<br />

have to make sure that these rules apply<br />

to every European market and to every<br />

market participant in Europe to the very<br />

same extent.<br />

Rainer Riess<br />

Michael Krogmann<br />

Deutsche Börse AG<br />

5 Paradigm Change Through<br />

Algorithmic Trading<br />

5.1 Introduction<br />

Algorithmic trading nowadays often accounts<br />

for more than half of trade volume<br />

and order volume at large stock<br />

exchanges. Its net effects are generally<br />

found positive by researchers. Only few<br />

voices from the scientific community –<br />

more,however,fromtraders–pointout<br />

negative effects of algorithmic trading. A<br />

notable difference lies between empirical<br />

findings – that usually find positive effects<br />

– on the one hand, and some theoretical<br />

works and especially the sentiment<br />

of traders, who often express their frustration<br />

about their computerized counterparts,<br />

on the other hand.<br />

5.2 Availability of Data<br />

Most scientific studies about algorithmic<br />

trading share one fundamental problem:<br />

data about algorithmic trading are<br />

scarce. As one of the few stock exchanges,<br />

Deutsche Börse had for some<br />

time quite reliable data on algorithmic<br />

trading. Their “Automated Trading<br />

Program“ (ATP), which was in effect<br />

from 2007 to early 2009, enabled them<br />

to distinguish between algorithmic orders<br />

and human ones (Deutsche Börse<br />

2009). Hendershott and Riordan (2011),<br />

Gsell (2009), Groth (2009), and Maurer<br />

and Schäfer (2011) analyze such datasets<br />

which contain flags for orders placed<br />

within the ATP environment. Their research<br />

questions differ, but they all more<br />

or less conclude that the overall effect of<br />

algorithmic trading is positive.<br />

A fundamental critique of such analyses<br />

is that algorithms usually work well<br />

in “normal” markets and then show the<br />

often-found positive effects. The models<br />

that algorithms base on are abstractions<br />

of reality and must fail to reflect it in its<br />

entirety. If a market situation is not part<br />

of the possibility space of the model, several<br />

options are possible: The algorithm<br />

halts trading and waits until the market is<br />

“normal” again, thereby facing the risk to<br />

generate possibly considerable losses. Another<br />

option is to continue trading using<br />

the usual model, thus failing to trade optimally<br />

and possibly worsening the situation.<br />

Since the flash crash on May 6, 2010,<br />

there have been repeated miniature flash<br />

crashes that did not affect the whole market<br />

but only individual stocks. For both<br />

phenomena, algorithms are blamed to be<br />

the cause of the market irregularities.<br />

However, an effective approach to regulation<br />

should base on well-established<br />

results. A lot of work has to be done<br />

here. Above all, the insufficient availability<br />

of appropriate data confines scientific<br />

progress. The deduction of the effect<br />

of algorithmic trading on the market<br />

from anonymous order book data can<br />

only be very rudimentary. In our current<br />

work, we attempt to find a way to analyze<br />

algorithmic trading activity whilst<br />

only using anonymous order book data<br />

98 Business & Information Systems Engineering 2|2012


(Breuer and Burghof 2011). A mandatory<br />

flagging of algorithmic orders would<br />

be desirable. Only then would it be possible<br />

to independently analyze algorithmic<br />

trading from many points of view<br />

and estimate the effect on the market.<br />

The restrictive handling of historic ATP<br />

data by Deutsche Börse does not build<br />

confidence but could increase the probability<br />

that the sentiment towards AT is<br />

influenced by irrational fears.<br />

5.3 Information Efficiency<br />

Recent studies (Hendershott and Riordan<br />

2011; Gsell 2009; Groth 2009; Maurer<br />

and Schäfer 2011) analyze rather shortterm<br />

aspects of market microstructure in<br />

an AT environment. Indeed, its existence<br />

alters behavioral incentives of other market<br />

participants fundamentally and in the<br />

long run. It is apparent that algorithms<br />

process new information ever faster and<br />

– assuming normal market conditions –<br />

probably calculate its price impact better<br />

than humans. It is still to be seen, however,<br />

how accurate trading algorithms<br />

process information without slow human<br />

traders monitoring them. Sometimes, the<br />

superfast processing of news can be undesirable.<br />

An example for this is the news<br />

about the bankruptcy of United Airways.<br />

The airline’s stock price plummeted until<br />

it became clear that the news was already<br />

a couple months old. Because the<br />

possibility to extract yields from new information<br />

has a very short and decreasing<br />

half-life, systems tend to react hastily<br />

and without challenging the information.<br />

Especially in delicate market situations,<br />

rumors can develop a destructive power.<br />

The effect that is likely to be most<br />

important has however escaped scientific<br />

analysis so far. Capital markets are a<br />

highly efficient instrument of capital allocation,<br />

especially because a large number<br />

of actors feed information into the<br />

price via their trading activity. This information<br />

comes from various sources;<br />

it may be obtained haphazardly or with<br />

some effort. Algorithmic trading uncovers<br />

trade activity which is caused by that<br />

information and uses this knowledge to<br />

pocket a considerable part of the information<br />

yield. The better these algorithms<br />

work, the less money the informed person<br />

will make out of this information. In<br />

the long run, this could mean that the<br />

costly generation of information turns<br />

unprofitable, and in an extreme case even<br />

the trade based on incidentally obtained<br />

information does not pay anymore.<br />

In such a hypothetical market, ever less<br />

information is traded ever more perfectly<br />

and faster. The market draws nearer and<br />

nearer the weak form of market efficiency<br />

(Fama 1970) or eventually even the semistrong<br />

form of market efficiency. At the<br />

same time, it moves away from the strong<br />

form of market efficiency, because the incentive<br />

to feed information into the market<br />

becomes considerably less powerful.<br />

It is this very effect that traders witness<br />

when they trade against algorithms. They<br />

know that information-based strategies<br />

are detected rapidly and thwarted by appropriate<br />

front-running strategies (Biais<br />

et al. 2010; Cvitanic and Kirilenko 2010).<br />

Surly, there is still a need for theoretical<br />

as well as empirical analysis here, because<br />

duetothesethoughts,theusefulnessof<br />

algorithmic trading is subject to scrutiny.<br />

5.4 Regulation and Regulatory<br />

Instruments<br />

Regulatory considerations have to distinguish<br />

between the different types of algorithms.<br />

Limit orders which are bogus orders<br />

or part of quote-stuffing techniques<br />

have to be considered under the light of<br />

laws against market manipulation (e.g.,<br />

§20a (1) No. 2 of the German Securities<br />

Trading Act [WpHG]). Other strategies<br />

improve the price quality by arbitraging<br />

prices and equalizing them across different<br />

trading venues. Because of the market<br />

power of algorithms, there is the risk<br />

that overly mechanic thinking and potent<br />

algorithms may perturb the price formation<br />

process. Naturally it would be desirable<br />

to capture the positive effects of<br />

algorithmic trading and to dampen the<br />

potentially negative ones. There may be<br />

more than one way to reach this aim.<br />

A simple ban of algorithmic trading, as<br />

sometimes demanded by certain political<br />

circles, cannot serve to reach this difficult<br />

aim. This would mean to also destroy<br />

many preferable effects of algorithmic<br />

trading. Of course, a distinction of<br />

algorithmic and “normal” trading is not<br />

easy. And certainly market participants<br />

would program algorithms that operate<br />

inthegrayareatohidetheirtruenature.<br />

Currently, regulatory bodies are discussing<br />

possible means (Dombert 2011).<br />

The often contemplated lower boundary<br />

for limit order lifetimes is regarded sceptically.<br />

The comprehensible reason is that<br />

an efficient risk management of orders<br />

would be drastically complicated – especially,<br />

but not exclusively, in volatile<br />

BISE – DISCUSSION<br />

markets. Dombert (2011)proposesanalternative<br />

that is worth discussing. With<br />

an order-transaction-ratio, the number<br />

of orders divided by the number of transactions<br />

would have to remain above some<br />

exogenous constant.<br />

In our view, a European regulatory<br />

framework is desirable that defines the<br />

playground for all market participants.<br />

Within this framework, it should be left<br />

to the trading venues how they wish to<br />

treat algorithmic trading in the context<br />

of their business model. Then it would<br />

be up to them if they wanted to attract<br />

algorithmic trading or to limit it in specific<br />

market conditions. Such a “menuapproach”<br />

leaves it in essence to the individual<br />

trader if he or she wishes to<br />

face the competition from algorithms<br />

with all their positive and negative effects<br />

or evade them by trading on trading<br />

venues with appropriate restrictions that<br />

apply always or under specific market<br />

conditions.<br />

5.5 Conclusion<br />

As long as algorithms operate in the<br />

dark, there is a profound uncertainty<br />

about the effect of their activities. Therefore,<br />

algorithmic trading is partly in contradiction<br />

to fundamental principles of<br />

stock exchanges: bringing buyers and<br />

sellers together in a transparent manner.<br />

On stock exchanges, trust is paramount.<br />

The opacity of algorithmic trading –<br />

as comprehensible it may be from the<br />

point of view of their operators – undermines<br />

this principle. Currently, there<br />

is no level playing field. However, it<br />

is equally important to enable technical<br />

progress, which algorithmic trading<br />

with its high-quality information processing<br />

definitely is. An improved availability<br />

of data and associated scientific<br />

research can help to develop reasonable<br />

regulatory frameworks for algorithmic<br />

trading. With the increasing importance<br />

of this way of trading in mind, there is<br />

less and less reason to doubt that the implementation<br />

of appropriate regulatory<br />

frameworks should have a high priority.<br />

Arne Breuer<br />

<strong>Prof</strong>. <strong>Dr</strong>. Hans-Peter Burghof<br />

University Hohenheim<br />

6 High Frequency Trading –<br />

ACentralBankView<br />

The capital markets are currently at an<br />

important juncture in their development.<br />

Business & Information Systems Engineering 2|2012 99


BISE – DISCUSSION<br />

Roughly half of all stock and foreign exchange<br />

trades conducted on the major<br />

exchanges are no longer initiated by human<br />

traders; instead, they are the product<br />

of computer algorithms that are able<br />

to analyze large volumes of data and initiatehundredsofordersinfractionsof<br />

a second. Humans are increasingly being<br />

eliminated from the immediate decisionmaking<br />

process relating to the sale and<br />

purchase of assets and being replaced by<br />

software programs.<br />

The speed with which orders are executed<br />

has become to be the most important<br />

factor and is now measured in milliand<br />

microseconds. New practices such as<br />

“co-location” or “quote stuffing” – placing<br />

huge quantities of buy or sell orders<br />

which the instigator intends to cancel almost<br />

immediately before they are executed<br />

– have become important instruments<br />

in the battle for the most rapid order<br />

execution. Fundamental data on the<br />

value of the respective securities or currencies<br />

are of no, or only subordinate,<br />

importance for HFT algorithms.<br />

In HFT, positions are usually held for<br />

between a number of milliseconds and<br />

several hours. In today’s high-speed markets,<br />

the scales are no longer tipped in<br />

favor of the investor who is best able to<br />

assess the true value of an asset, but of<br />

the investor able to trade fastest. True<br />

investments are becoming increasingly<br />

rare.<br />

Since the “flash crash” of May 6, 2010<br />

(a roughly 15-minute phase of unusual<br />

and irrational volatility on the New York<br />

Stock Exchange), HFT has been called<br />

to the attention not only of the general<br />

public but also of regulators and central<br />

banks.<br />

Numerous observers regard HFT as a<br />

new technical means of executing existing<br />

trading strategy rather than a strategy<br />

in its own right. Advantages in terms<br />

of speed have, they say, always been an<br />

essential component of many successful<br />

trading strategies. Seen from this perspective,<br />

HFT is not a completely new<br />

phenomenon, but rather a technical evolution<br />

of the securities markets. HFT<br />

should be regarded merely as an overarching<br />

term covering a multitude of different<br />

fields of use. Among the many tactics,<br />

several of the most important are based<br />

on providing liquidity in stock market<br />

trading (market making). Others can be<br />

included under the category “statistical<br />

arbitrage” and use algorithms to swiftly<br />

identify and exploit profitable trading<br />

opportunities based on price data. Others<br />

belong to a category known as liquidity<br />

detection, in which traders try to<br />

seek out hidden large orders in order<br />

books. Many critics term this “predatory<br />

trading”, and it is suspected of being<br />

unfair and potentially damaging to the<br />

market.<br />

Against this complex background, any<br />

assessment of HFT and all discussion<br />

relating to potential regulation should,<br />

where possible, be limited to the underlying<br />

HFT strategy. From a central<br />

bank perspective, a sweeping judgment<br />

on HFT is therefore neither appropriate,<br />

nor would it serve to improve the<br />

regulatory framework for transparency,<br />

stability and efficiency. That means that<br />

both the advantages and disadvantages of<br />

HFT need to be evaluated very specifically.<br />

Statements that HFT is in general<br />

either good or bad for the market should<br />

therefore be viewed with caution.<br />

HFT players and exchange operators<br />

are at pains to stress that overall HFT<br />

perceptibly improves market liquidity<br />

and the efficiency of price discovery<br />

(McEachern Gibbs 2009). The majority<br />

of investors benefit from reduced bid/ask<br />

spreads – a common measure of liquidity,<br />

they say. This statement is backed<br />

up by several scientific studies (Gomber<br />

et al. 2011). However, there is increasing<br />

evidence to suggest that, especially<br />

in very volatile market situations, HFT<br />

could prove problematic and could additionally<br />

destabilize the market (Brogaard<br />

2010). This must be investigated and, if<br />

found to be true, regulators must step in<br />

to limit the risks for the financial system.<br />

The flash crash demonstrated that the<br />

liquidity generated by HFT market makers,<br />

which usually keeps transaction costs<br />

low, may suddenly evaporate in difficult<br />

market phases (NANEX 2010). Unlike<br />

regular “human” market makers, who are<br />

obliged to remain in the market even in<br />

times of extremely volatile prices, HFT<br />

tradersaregenerallynotboundbysuch<br />

constraints. In good times, HFT traders<br />

therefore crowd out normal market makers<br />

and often even perform their role better,<br />

to the advantage of all market players.<br />

In difficult markets, however, there is a<br />

risk that trading flows could collapse with<br />

all the attendant problems for the market<br />

as a whole, as HFT players withdraw.<br />

To many market participants, the narrower<br />

bid/ask spreads and higher trading<br />

volume generated by HFT therefore<br />

only represent “sham liquidity”. For this<br />

reason there have been calls from various<br />

quarters to oblige HFT market makers to<br />

remain in the market even in times of<br />

high volatility, similar to the obligations<br />

imposed on normal market makers (EC<br />

2010). In other words, they should start<br />

to take some responsibility for the markets<br />

which they have, to date, merely used<br />

to their advantage from their superior<br />

position.<br />

From a regulatory perspective, HFT<br />

has proven problematic not only in these<br />

rare but dramatic high volatility events,<br />

but also in daily trading activities. While<br />

bid/ask spreads have dropped significantly<br />

in recent years thanks to HFT<br />

market makers, the average period for<br />

which such players hold positions has<br />

dropped sharply. According to a study on<br />

the flash crash, most HFT market makers<br />

close out their positions after no more<br />

than roughly 10 seconds (Kirilenko et<br />

al. 2011). That means that the stabilizing<br />

effect in the event of heightened market<br />

volatility exerted by “normal” market<br />

makers has given way to a “hot potato<br />

effect”, where falling shares are merely<br />

passed around at lightning speed.<br />

As HFT has become more widespread,<br />

the number of buy and sell orders has increased<br />

dramatically in recent years. The<br />

tactic known as quote stuffing, which is<br />

used by several HFT algorithms, is particularly<br />

problematic. For reasons of tradingstrategy,theHFtraderplacesalarge<br />

number of orders per second, only to<br />

cancel them again almost immediately<br />

before execution. The very high cancellation<br />

rate this causes leads to a marked<br />

divergence between apparent market liquidity<br />

and actual trading volume. An investor<br />

placing an order in response to<br />

a bid or ask is therefore often unable<br />

to carry out the transaction at the limit<br />

shown. Although the explicit transaction<br />

costs appear low, the implied costs may<br />

be much higher. Apparent market liquidity<br />

and the size of bid/ask spreads<br />

are therefore not by themselves reliable<br />

indicators of market liquidity and<br />

efficiency.<br />

An analysis of 1,172 trading days on<br />

the New York Stock Exchange from 2007-<br />

01-01 to 2011-09-14 that was carried out<br />

recently by the research firm NANEX<br />

showed that there were just 35 billion real<br />

transactions for 535 billion quotes. The<br />

quotes-to-trades ratio needed to generate<br />

US$ 10,000 in real transaction volume<br />

moved from roughly 6–7 at the beginning<br />

of 2007 to 60–80 in mid-2011.<br />

Higher figures indicate a less efficient<br />

market: more information is required to<br />

100 Business & Information Systems Engineering 2|2012


achieve the same trading volume. Sudden<br />

and dramatic spikes in the number<br />

of quotes are increasingly being observed<br />

for individual US stocks, with individual<br />

HFT algorithms generating several<br />

tens of thousands of quotes per second<br />

for several seconds. Such bursts of activity<br />

are frequently accompanied by what<br />

are known as “mini flash crashes”, where<br />

securities lose 20%, 40% or even more<br />

than 50% of their value in a space of seconds<br />

for no fundamental reason, only to<br />

recover shortly afterwards. For instance,<br />

according to the SEC, the United States<br />

has witnessed more than 100 such inexplicable<br />

crashes since mid-2010 which<br />

are suspected of being caused by HFT<br />

algorithms.<br />

Sending bids or asks is similar to sending<br />

spam email: both are virtually free for<br />

the sender, but not for the recipient. Forwarding<br />

and processing such large volumes<br />

of data causes a lot of problems<br />

and high costs for exchanges and market<br />

participants. Systems are often overloaded,<br />

which is seen by many observers<br />

as one of the causes of the flash crash. To<br />

make matters worse, certain HFT algorithms<br />

send some of these quotes only to<br />

cause other traders or algorithms to act in<br />

a certain way, which they can, in turn, exploit.<br />

As a consequence, an ever increasing<br />

number of institutional investors are<br />

transferring their transactions away from<br />

normal exchanges to “dark pools”, where<br />

it is usually more difficult to make a profit<br />

in HFT.<br />

The above-described criticisms intend<br />

to show that HFT is a controversial issue,<br />

requiring an exact analysis of the<br />

details. In addition to “market friendly”<br />

strategies that regulators regard as positive<br />

for the market – for instance, statistical<br />

arbitrage – there are also “unfriendly”<br />

strategies that are seen as worrying. Others<br />

are basically welcome but when actually<br />

applied on the market entail problems<br />

and dangers which should be eliminated.<br />

HFT market making is just such<br />

an example.<br />

When considering the ultimate question<br />

of whether there is a correlation<br />

between HFT and market efficiency, it<br />

should be borne in mind that market efficiency<br />

mainly means that the price of<br />

an asset adjusts to fundamental changes<br />

in its value rapidly. It is not immediately<br />

clear how HFT algorithms can contribute<br />

to that, as decisions are based only on<br />

the status of the order book in the last<br />

few seconds or indicators based on technical<br />

analysis. A block trade of 10,000<br />

shares between two well-informed large<br />

investors represents true price discovery<br />

on the market. By contrast, shifting 100<br />

shares back and forth between two HFT<br />

algorithms in innumerous times makes<br />

no equivalent contribution to trading efficiency,<br />

even if this takes place at impressive<br />

speed. A market that is mainly<br />

dominated by HFT is also a market where<br />

most orders have lost all connection to<br />

fundamental factors. And this correlation<br />

between price and fundamental value is<br />

what should, in the main, determine the<br />

quality of a market.<br />

References<br />

To: Section 1<br />

<strong>Dr</strong>. Joachim Nagel<br />

<strong>Dr</strong>. Rafael Zajonz<br />

Deutsche Bundesbank<br />

Aldridge I (2009) High-frequency trading.<br />

A practical guide to algorithmic strategies.<br />

Wiley, Hoboken<br />

Brogaard JA (2010) High frequency trading<br />

and volatility. Working paper, Northwestern<br />

University, Chicago<br />

Dacorogna MM, Gencay R, Müller U, Olsen<br />

RB, Pictet OV (2001) An introduction to<br />

high-frequency finance. Academic Press,<br />

San Diego<br />

Durbin M (2010) All about high frequency<br />

trading – the easy way to get started.<br />

McGraw Hill, New York<br />

Gomber P, Arndt B, Lutat M, Uhle T (2011)<br />

High-frequency trading. Working paper,<br />

Goethe-Universität Frankfurt<br />

Gomolka J (2011) Algorithmic Trading: Analyse<br />

von computergesteuerten Prozessen<br />

im Wertpapierhandel unter Verwendung<br />

der Multifaktorenregression. Dissertation,<br />

Potsdam, Universitätsverlag Potsdam<br />

Hendershott T, Riordan R (2011) Algorithmic<br />

trading and information. Working paper,<br />

University of California (Berkeley), Karlsruhe<br />

Institute of Technology<br />

Ye G (2011) High frequency trading models +<br />

website. Wiley, Hoboken<br />

Zhang F (2010) High-frequency trading,<br />

stock volatility, and price discovery. http://<br />

ssrn.com/abstract=1691679. Accessed<br />

2012-01-30<br />

To: Section 2<br />

European Commission (2010) European<br />

Commission directorate general internal<br />

market and services. Public consultation:<br />

review of the markets in financial<br />

instruments directive (MiFID). http://ec.<br />

europa.eu/internal_market/consultations/<br />

docs/2010/mifid/consultation_paper_en.<br />

pdf. Accessed 2011-01-19<br />

Gomber P, Arndt B, Lutat M, Uhle T (2011)<br />

High frequency trading. Working paper,<br />

Universität Frankfurt<br />

Kirilenko A, Kyle A, Samadi M, Tuzun T (2011)<br />

The flash crash: the impact of high frequency<br />

trading on an electronic market.<br />

Working paper<br />

To: Section 3<br />

BISE – DISCUSSION<br />

Brogaard JA (2010) High frequency trading<br />

and volatility. Working paper, Northwestern<br />

University, Chicago<br />

Forbes (2010) Wall Street’s speed war. 2010-<br />

09-27<br />

Hendershott T, Jones CM, Menkveld AJ (2011)<br />

Does algorithmic trading improve liquidity?<br />

J <strong>Finance</strong> 66(1):1–33<br />

Hendershott T, Riordan R (2011) High frequency<br />

trading and price discovery. Working<br />

paper, University of California (Berkeley),<br />

Karlsruhe Institute of Technology<br />

Securities and Exchange Commission (SEC)<br />

(2010) Concept release on equity market<br />

structure. Release No. 34-61358, File<br />

No. S7-02-10<br />

Zhang S, Riordan R (2011) Technology and<br />

market quality: the case of high frequency<br />

trading. In: Proc ECIS 2011<br />

To: Section 4<br />

Deutsche Börse AG (2011) Several internal papers<br />

and statistics. August 2011, Eschborn<br />

Brogaard JA (2010) High frequency trading<br />

and volatility. Working paper, Northwestern<br />

University, Chicago<br />

AFM/Netherlands Authority for the Financial<br />

Markets (2010) High frequency trading<br />

– the application of advanced trading<br />

technology in the European marketplace<br />

SEC and CFTC (2010) Findings regarding the<br />

market events of May 6, 2010<br />

Zhang S, Riordan R (2011) Technology and<br />

market quality: the case of high frequency<br />

trading. In: Proc ECIS 2011<br />

To: Section 5<br />

Biais B, Foucault T, Moinas S (2010) Equilibrium<br />

algorithmic trading. Working paper,<br />

Toulouse School of Economics<br />

Breuer A, Burghof H-P (2011) An empirical<br />

analysis of order dynamics in a highfrequency<br />

trading environment. Unpublished<br />

working paper, Universität Hohenheim<br />

Cvitanic J, Kirilenko A (2010) High frequency<br />

traders and asset prices. Unpublished<br />

working paper, CFTC<br />

Deutsche Börse (2009) Price list for the utilization<br />

of the trading system of FWB Frankfurt<br />

Stock Exchange. Deutsche Börse AG<br />

Dombert AR (2011) Hochfrequenzhandel:<br />

Nutzen und Risiken neuer Technologien<br />

auf den Finanzmärkten. Kreditwesen 18:<br />

16–18<br />

Fama E (1970) Efficient capital markets: a review<br />

of theory and empirical work. J <strong>Finance</strong><br />

25(2):383–417<br />

Gsell M (2009) Algorithmic activity on Xetra.<br />

J Trading 4(3):74–86<br />

Groth S (2009) Algorithmic trading engines<br />

and liquidity contribution: the blurring<br />

of “traditional” definitions. In: Godart C,<br />

Gronau N, Sharma S, Canals S (eds) Proc 9th<br />

IFIP conference on e-business, e-services,<br />

and e-society (I3E 2009)<br />

Hendershott T, Riordan R (2011) High frequency<br />

trading and price discovery. Working<br />

paper, University of California (Berkeley),<br />

Karlsruhe Institute of Technology<br />

Business & Information Systems Engineering 2|2012 101


BISE – DISCUSSION<br />

Maurer K-O, Schäfer C (2011) Analysis of binary<br />

trading patterns in Xetra. J Trading<br />

6(1):46–60<br />

To: Section 6<br />

McEachern Gibbs C (2009) HFT is market<br />

making evolved, says NYSE exec.<br />

http://www.advancedtrading.com/<br />

managingthedesk/219501268. Accessed<br />

2012-01-30<br />

Gomber P, Arndt B, Lutat M, Uhle T (2011)<br />

High frequency trading. Working paper,<br />

Universität Frankfurt<br />

Brogaard JA (2010) High frequency trading<br />

and volatility. Working paper, Northwestern<br />

University, Chicago<br />

NANEX (2010) Analysis of the “flash<br />

crash”. http://www.nanex.net/20100506/<br />

FlashCrashAnalysis_CompleteText.html.<br />

Accessed 2012-01-30<br />

European Commission (2010) European<br />

Commission directorate general internal<br />

market and services. Public consultation:<br />

review of the markets in financial<br />

instruments directive (MiFID). http://ec.<br />

europa.eu/internal_market/consultations/<br />

docs/2010/mifid/consultation_paper_en.<br />

pdf. Accessed 2011-01-19<br />

Kirilenko A, Kyle A, Samadi M, Tuzun T (2011)<br />

The flash crash: the impact of high frequency<br />

trading on an electronic market.<br />

Working paper<br />

102 Business & Information Systems Engineering 2|2012


Investigating the Market Impact of<br />

Media Sentiment and Investor Attention<br />

Michael Siering 1<br />

1 Goethe-University Frankfurt, Grüneburgplatz 1, 60323 Frankfurt, Germany,<br />

siering@wiwi.uni-frankfurt.de<br />

Abstract. Media sentiment has been shown to be related to stock returns.<br />

However, one prerequisite for this influence has not been taken into account<br />

yet: the question whether investors actually pay attention to news and the<br />

related financial instruments. Within this study, we close this research gap by<br />

examining the interplay between media sentiment and investor attention.<br />

Thereby, we find that the positive impact of media sentiment on returns is<br />

increased when investor attention is high. Furthermore, we evaluate whether<br />

these variables can be used to forecast future market movements. Although our<br />

results reveal that the obtained forecasting accuracy cannot be achieved by<br />

chance, we conclude that further information has to be included in the<br />

forecasting model to obtain satisfying results.<br />

Keywords: Media Sentiment, Investor Attention, Behavioral <strong>Finance</strong>.<br />

1 Introduction<br />

Within recent years, the impact of media sentiment on financial markets has been of<br />

great interest. Recent studies have found that sentiment expressed in traditional<br />

mainstream media like newspapers or social media such as blogs and message boards<br />

is related to stock returns [1, 2]. Additionally, the question whether investors are<br />

aware of financial news has also been examined in order to explain anomalies like the<br />

slow incorporation of new information into stock prices on Fridays [3]. Furthermore,<br />

the impact of an increased number of irrelevant news on the processing of relevant<br />

ones has been examined as well [4].<br />

However, to the best of our knowledge, previous research has neglected the<br />

interplay between investor attention and media sentiment as well as the following<br />

influence on financial markets. Nevertheless, it seems intuitive to take the interaction<br />

between both variables into account: the prerequisite for investors being influenced by<br />

media sentiment is that investors actually notice the news articles published within the<br />

media. Consequently, it can be assumed that a higher level of investor attention leads<br />

to an increased media sentiment impact on financial markets.<br />

Furthermore, the contemporaneous influence of media sentiment on financial<br />

markets still remains underexplored: Recent studies investigate the impact of media<br />

sentiment with at least a one day delay [2, 5]. In contrast, we show that many news<br />

articles are published during the day. Since new information contained within these


news articles is usually processed within short periods of time [6], it also seems<br />

reasonable to investigate the contemporaneous impact of media sentiment.<br />

We contribute to the literature in several ways. At first, we investigate the<br />

contemporaneous impact of media sentiment and investor attention on stock returns.<br />

Second, we investigate the interplay of media sentiment as well as investor attention<br />

and its impact on the following stock market reactions. Third, we forecast future<br />

returns taking into account those variables.<br />

Based on an analysis of the media sentiment related to the Dow Jones Industrial<br />

Average (DJIA), we confirm the impact of media sentiment on DJIA returns and<br />

enhance the previous understanding related to the fact that media sentiment already<br />

has an influence on financial markets at the same day when the corresponding news<br />

are published. Additionally, we find that positive media sentiment has an increased<br />

influence on DJIA returns when investor attention is high. Related to forecasting<br />

future DJIA returns, we conclude that the accuracy that can be obtained is higher than<br />

results that are achieved by chance. However, additional information has to be<br />

included in the forecasting model to obtain satisfying results.<br />

The remainder of this paper is structured as follows. At first, we present related<br />

work concerning the influence of media sentiment and investor attention on financial<br />

markets. Thereafter, we outline the data used within our study, derive our sentiment<br />

measure and describe the proxy for quantifying investor attention. Next, we consider<br />

the joint impact of media sentiment and investor attention on DJIA returns.<br />

Thereafter, we evaluate whether future market movements can be forecasted by<br />

taking into account these variables. Finally, we conclude this paper.<br />

2 Related Work<br />

2.1 Media Sentiment and Financial Markets<br />

Investors are considered to decide to trade because of two general reasons. First, they<br />

take into account new fundamental information like dividend announcements or<br />

management decisions [7]. Second, they are also considered to rely on expectations<br />

that do not follow rational rules [8]. For instance, these expectations can be based on<br />

the advice of “financial gurus” [9] or simply on the sentiment prevailing in the media<br />

that causes them to be overconfident in making the right decisions [10]. In this<br />

context, sentiment expressed in the media covers opinions, expectations or beliefs of<br />

market participants towards certain companies or towards certain financial<br />

instruments [11]. If many investors take media sentiment into account, have similar<br />

(irrational) expectations and follow each other, this can influence stock prices [12].<br />

Recent research has provided evidence for these assumptions. It has been shown that<br />

sentiment expressed in media has an impact on investors’ decision-making activities<br />

and thus affects several financial variables. Consequently, investors act according to<br />

their expectations and buy or sell the respective financial instrument.<br />

Different studies investigate the impact of sentiment expressed in traditional media<br />

like newspapers. Within this context, [5] analyzes a daily Wall Street Journal column<br />

and finds that high media pessimism leads to a decline in market prices. Additionally,


an abnormal high or low level of pessimism is supposed to predict high trading<br />

volumes. A similar study is conducted by [2]. On a daily basis, they analyze the news<br />

stories published in the Wall Street Journal as well as in the Dow Jones News Service<br />

and confirm that stock prices are related to media sentiment. In contrast, [13] evaluate<br />

the sentiment of corporate disclosures (i.e. 10-K company reports) and find that,<br />

compared to general approaches, domain-specific sentiment measures are more<br />

appropriate for sentiment detection within this document type. Furthermore, they<br />

confirm the relation of sentiment and several financial variables.<br />

Compared to these studies, another stream of research focuses on the impact of<br />

sentiment expressed in social media. A seminal study that investigates the impact of<br />

sentiment on a stock level is presented by [1], who collect and analyze messages<br />

posted on two finance message boards. The authors find that disagreement in<br />

sentiment among the messages leads to an increase in trading volume. Additionally,<br />

they observe that the number of messages posted during a day can help to predict the<br />

stock returns during the following day. [14] follow a similar approach and investigate<br />

messages which are published on stock message boards, too. However, they focus on<br />

an index rather than stock level. Thereby, they determine the sentiment for every<br />

message whereas these messages are then used to calculate an overall sentiment<br />

index. [14] find that the level of this sentiment index has explanatory power for the<br />

level of the corresponding stock index. In contrary to this result, they only find weak<br />

evidence that the sentiment concerning individual stocks can forecast daily stock price<br />

movements. Apart from these results, recent studies have found a link between the<br />

sentiment prevailing in microblogging services like twitter and financial markets [15].<br />

The studies presented above provide evidence that sentiment expressed in news<br />

articles or message board postings is related to stock returns. However, these studies<br />

mainly focus on the long-term effect of media sentiment on financial variables.<br />

Instead of taking into account contemporaneous effects, stock returns are related to<br />

the previous days’ media sentiment. However, due to the fact that new information is<br />

often processed within minutes rather than days [6] it seems possible that the related<br />

sentiment has a contemporaneous effect as well.<br />

2.2 Investor Attention and Financial Markets<br />

In recent years, several financial market anomalies have been investigated<br />

theoretically and empirically, such as underreaction and overreaction to financial<br />

news, the influence of weekdays on investors’ reactions as well as the impact of<br />

advertisements on investors’ decisions. Many of these anomalies have been attributed<br />

to the level of investor attention, i.e. the question whether investors are aware of the<br />

current market situation or not. Thus, there is a large number of studies investigating<br />

which instruments receive attention, how corporate advertising impacts the level of<br />

investor attention and how investors pay attention to news published by firms or by<br />

the media in general. Concerning the question which financial instruments are of<br />

interest for different groups of investors, [16] examine how individual and<br />

institutional investors react to “attention-grabbing” stocks. Thereby, they find that<br />

individual investors especially pay attention to stocks which are discussed within the<br />

media, exhibit high abnormal trading volumes and high returns.


Next to the question which financial instruments gain attention in general, another<br />

stream of research investigates how stock recommendations published within the<br />

media influence investor attention. Within this context, it has been found that trading<br />

volumes increase after a stock is discussed within television [17]. Additionally, it has<br />

been figured out that a firm’s advertising expenses lead to an increased number of<br />

individual investors buying a stock [18]. In this case, a spillover effect of<br />

advertisements can be measured: although a firm advertises its product and intends to<br />

increase the product related attention, there is also a higher interest related to its<br />

stocks. These results are confirmed by [19]. Furthermore, [19] find that<br />

advertisements lead to an increase in stock returns in the contemporary year, but they<br />

also note that stock returns decrease in the following year. Considering the company<br />

size, they find that this effect is larger for small firms [19].<br />

Next to stock recommendations and advertisements, investor attention is also<br />

influenced by ordinary news published within the media. In this context, [20] analyze<br />

the market reactions on news of economically linked firms. In this context, they find<br />

that news are incorporated slower when they are not directly related to the firm under<br />

investigation but deal with an economically linked firm. This effect is attributed to a<br />

small degree of investor attention. Additionally, a study by [4] investigates whether<br />

the amount of news articles published within the same period of time has an impact<br />

on market reactions. [4] find that an increased amount of unimportant news decreases<br />

the investors’ reactions to relevant news. Thereby, price and volume reactions are<br />

lower and the post-announcement adjustment to the news is stronger. Thus, as a<br />

result, an increased number of news articles published in the same period is said to<br />

reduce investor attention towards specific news items. Considering investor attention<br />

on different days of the week, [3] find that the response to earnings announcements is<br />

slower on Fridays compared to the remaining days of the week. This effect is<br />

attributed to a lower level of investor attention on Fridays, whereas investors are said<br />

to be more distracted because of the following weekend. Similar results concerning<br />

limited investor attention on Fridays are also reported by [21]. Additionally, [22]<br />

provide evidence that the level of investor attention is related to stock returns.<br />

Based on these studies, it can be noted that the level of investor attention has an<br />

influence on the market reactions following the publication of financial news. In this<br />

context, investors who are aware of the news articles published are confronted with<br />

the corresponding sentiment. As follows, it is more likely that their expectations and<br />

trading decisions are influenced by media sentiment. Thus, a joint effect of both<br />

variables on financial markets can be expected. However, previous studies have not<br />

focused on this specific relation of media sentiment and investor attention.<br />

3 Research Methodology<br />

3.1 Measuring Media Sentiment<br />

In general, sentiment analysis encompasses the investigation of documents like news<br />

articles, message board postings or product reviews in order to determine their tone<br />

concerning a certain topic [23, 24]. There are two broad strategies to perform


sentiment analysis: it can be distinguished between supervised and unsupervised<br />

approaches [25]. Supervised approaches require a dataset composed of documents<br />

that are manually labeled according to the respective sentiment. After several preprocessing<br />

steps, this dataset is used to train machine learning classifiers like naive<br />

bayes or support vector machines. During the training phase, the classifiers search for<br />

patterns within the documents. These patterns can thereafter be used to determine the<br />

sentiment of further documents or sentences. In contrast, unsupervised approaches<br />

rely on external knowledge such as predefined dictionaries providing lists of words<br />

that are connected with a positive or negative sentiment. These word lists are usually<br />

created manually with a couple of precoded terms and are used to determine a<br />

sentiment measure [26].<br />

Within our study, we decide to follow an unsupervised dictionary-based approach<br />

which determines the sentiment taking into account a dictionary containing sentiment<br />

bearing words [26]. This is appropriate because dictionary-based approaches have<br />

proven to be very promising within the financial domain [2, 5, 13]. In contrast,<br />

applying a supervised machine learning-based approach would require a manually<br />

labeled dataset for training whereas manual labeling would be time-consuming and<br />

error-prone.<br />

For unsupervised approaches, different dictionaries are available that contain<br />

positive and negative expressions. Within this study, we make use of the Harvard-IV-<br />

4 dictionary. This dictionary has often been applied in the financial context [2, 5].<br />

Since we analyze general financial news articles rather than specific corporate<br />

disclosures, we make use of this dictionary instead of using the specific dictionary<br />

proposed by [13] which was suggested for the analysis of corporate disclosures.<br />

To calculate a daily sentiment index, we first determine the sentiment of each<br />

document. Accordingly, we obtain the occurrences of positive and negative words by<br />

comparing each news article with the positive and negative word lists. To take<br />

negations into account, we follow [13] and reverse the interpretation of a word if it is<br />

preceded by a negation so that positive words are counted as negative and vice versa.<br />

Thereafter, we adapt a document-level sentiment polarity measure which determines<br />

the direction of the sentiment (i.e. ranging from negative to positive) as well as its<br />

strength [2, 27]:<br />

The measure defined in equation 1 takes into account the number of positive words<br />

pos as well as the number of negative words neg, calculated as described above. If a<br />

document contains neither positive nor negative words, sentdoc is defined as zero. In<br />

line with [2], this measure assumes that all positive and negative words are equally<br />

important, i.e. no weights are assigned to certain words.<br />

As a next step, we determine a daily sentiment index by aggregating the documentlevel<br />

sentiment on a daily basis. Therefore, we calculate the average of sentdoc. In the<br />

following, the resulting daily sentiment index sent that takes into account the<br />

sentiment related to the DJIA is used to investigate the research questions of our<br />

study.<br />

(1)


3.2 Measuring Investor Attention<br />

Within previous studies, different approaches for measuring investor attention have<br />

been proposed. In general, it can be distinguished between indirect and direct<br />

measures of investor attention. On the one hand, a large body of literature deals with<br />

indirect measures of investor attention. Exemplary proxies used are unusual trading<br />

volumes or returns as well as the number of news articles published per day [4, 16,<br />

19, 20]. In this case, it is assumed that large trading volumes or extreme returns<br />

indicate that investors are extremely aware of a stock and respond more timely to new<br />

information, i.e. they trade this stock. As follows, these measures can be denoted as<br />

ex post measures of investor attention. In contrast, the number of news articles per<br />

day can be seen as an ex ante measure of investor attention: an increased amount of<br />

news articles is assumed to lead to an increased amount of investor attention related to<br />

the corresponding financial instruments [16]. However, an increase in these indirect<br />

measures only expresses the results of investors buying or selling a stock (ex post<br />

measures) or a general increase in media attention (ex ante measures). In contrast,<br />

these indirect measures do not indicate whether investors are interested in a financial<br />

instrument or whether the news articles in the media are actually noticed by them at<br />

all [16].<br />

In consequence, [22] provide a seminal paper about the direct measurement of<br />

investor attention, i.e. the measurement of investor attention without relying on tradebased<br />

proxies or the number of news articles published. Instead, they propose to take<br />

into account the amount of web searches related to the company under investigation,<br />

assuming that investors being interested in a financial instrument also search for<br />

related information. In this case, [22] make use of Google’s search volume index<br />

(SVI). Thereby, they find that this measure is correlated with indirect proxies for<br />

investor attention but that it encompasses investor attention in a more timely way.<br />

Furthermore, they note that this measure is especially suited to cover retail investor<br />

attention [22]. Since studies from other domains have already proven the applicability<br />

of the amount of search queries to forecast housing sales, car sales or the outbreak of<br />

influenza (e.g. [28]), we also decide to use SVI as a direct measure for investor<br />

attention.<br />

The SVI can be obtained via Google Trends for different search terms and for<br />

different time horizons (beginning with January 2004). However, SVI is only<br />

displayed for search terms that received attention exceeding a certain (unknown)<br />

threshold. Thus, identifying the correct search term to cover investor attention can be<br />

seen as a crucial step. Within our study, we decide to take the SVI related to the<br />

search term “DJIA” into account to represent investor attention related to the DJIA.<br />

An alternative would have been to download the SVI for each constituent of the DJIA<br />

separately. However, in this case, several problems would arise: As already noted by<br />

[22], some company names are ambiguous (e.g. searching for Kraft). Using ticker<br />

symbols instead could be an alternative, however, there are also some pitfalls in this<br />

case. At first, SVI is not available for every ticker symbol and second, some ticker<br />

symbols are ambiguous, too. For example, searching for “T” as ticker symbol for<br />

AT&T also leads to results related to T-Mobile, “HD” for Home Depot could also be<br />

interpreted as a search for the technical abbreviation “high definition” (as in HD-<br />

DVD) and the same applies to “BA” (Bank of America) which can also be an


abbreviation for British Airways. In these cases, the SVI would not cover the<br />

corresponding ticker symbol and would be inappropriate to measure investor<br />

attention. Thus, we decide to use the SVI for “DJIA” as a proxy for DJIA investor<br />

attention. In contrast to [22], who make use of the weekly SVI, we take into account<br />

the daily SVI in order to measure contemporaneous effects.<br />

3.3 Dataset Acquisition<br />

Within this study, we consider three data sources. First, we acquire financial news<br />

articles in order to determine the media sentiment index. Therefore, we make use of<br />

news articles published by Dow Jones Newswires (DJNS). Second, we download the<br />

SVI related to the DJIA from Google trends. Third, we acquire the corresponding<br />

DJIA closing prices and trading volumes from Yahoo! <strong>Finance</strong>.<br />

The news articles by DJNS are accessed via the application programming interface<br />

provided by Interactive Data. Thereby, we search for all news articles that are tagged<br />

by DJNS to deal with the constituents of the DJIA. We see DJNS as a representative<br />

source for financial news since DJNS is a major news provider that publishes<br />

financial news throughout the day and whose news are accessed by a large audience<br />

[5]. As revealed by a manual review of the news articles at hand, the assigned labels<br />

are too broad: news articles are already tagged to be related to a certain company<br />

when they mostly deal with its competitors. Thus, we only include those news articles<br />

into our analysis that contain the corresponding search term within the headline. Due<br />

to licence terms, we were able to request all news articles from 2011/01/01 until<br />

2012/02/29, so that 292 trading days could be analyzed. In total, the news article<br />

dataset obtained for this study consists of 13,696 news articles. Thereby, the dataset<br />

consists of different news categories. First, 6,454 regular financial news articles are<br />

included. Second, 7,176 news articles are included that explicitly deal with corporate<br />

disclosures and press releases. Finally, there are 66 news articles included in the<br />

dataset that contain analyst opinions. Thus, our news article dataset covers the full<br />

spectrum of articles that is available within a regular financial newspaper.<br />

As already discussed above, some ticker symbols and company names are<br />

ambiguous. As a result, the SVI cannot be acquired with an adequate accuracy for<br />

each ticker symbol separately. Thus, we decided to acquire the daily SVI for the<br />

search term “DJIA” from Google trends to measure investor attention. In this context,<br />

the SVI can be downloaded relative to the beginning of the corresponding month or<br />

relative to the beginning of the year 2004. Since the first option does not allow to<br />

compare the SVI across several months, we have chosen to download the SVI for our<br />

sample period relative to the search volume in 2004. Finally, we acquire the DJIA<br />

prices and trading volumes for the sample period from Yahoo! <strong>Finance</strong>.


4 Empirical Results<br />

4.1 Descriptive Results<br />

At first, we consider the daily number of news articles related to the DJIA and its<br />

constituents published by Dow Jones Newswires. Taking into account the daily<br />

distribution as indicated in Fig. 1, it first can be noted that the number of news articles<br />

published per day is not constant over time. Instead, a much smaller amount of news<br />

articles is published on weekends as compared to the rest of the week.<br />

Number of News Articles<br />

3500<br />

3000<br />

2500<br />

2000<br />

1500<br />

1000<br />

500<br />

0<br />

Mon Tue Wed Thu Fri Sat Sun<br />

Day<br />

Fig. 1. Total Number of News Articles Published per Day / per Hour<br />

Second, it can be observed that the number of news articles published from Monday<br />

to Thursday is relatively constant, except from a peak on Tuesday. It is notable that<br />

the number of news articles on Friday is smaller than during the remaining trade days.<br />

This result may be attributed to the fact that the general number of financial news<br />

issued by firms is smaller on Fridays, as already reported by [3], [29] as well as [30].<br />

Thereby, these related studies find that on Fridays, a lower fraction of earnings<br />

announcements is published. As follows, the number of news articles published<br />

dealing with these events is smaller, too. Another explanation could be that next to<br />

investor inattention, also journalists are distracted on Fridays because of the following<br />

weekend.<br />

Next, considering the time of day when news articles are published, it can be noted<br />

that a large peak can be found in the morning at the start of the trading hours (all<br />

times reported in Eastern Standard Time) and a small peak can be found at the end of<br />

the trading hours. Since the news articles published by Dow Jones Newswires are<br />

delivered electronically, a lot of information (and related sentiment) is released during<br />

the day after the newspapers have been printed in the morning.<br />

Furthermore, we consider the amount of Google searches for the Keyword “DJIA”.<br />

In this case, we find abnormal high search volumes in August 2011 (see Fig. 2).<br />

These high levels occur simultaneously with a decline of the DJIA caused by weak<br />

economic perspectives. Thus, we control for this abnormal movement within our<br />

further analyses. Additionally, Fig. 3 shows that investor attention is low on<br />

weekends. Thus, investors are distracted on Saturdays and Sundays [3].<br />

Number of News Articles<br />

1400<br />

1200<br />

1000<br />

800<br />

600<br />

400<br />

200<br />

0<br />

1 AM 3 AM 5 AM 7 AM 9 AM 11 AM 1 PM 3 PM 5 PM 7 PM 9 PM 11 PM<br />

Time


SVI<br />

25<br />

20<br />

15<br />

10<br />

5<br />

0<br />

20110101<br />

20110116<br />

20110131<br />

20110215<br />

20110302<br />

20110317<br />

20110401<br />

20110416<br />

20110501<br />

20110516<br />

20110531<br />

20110615<br />

20110630<br />

20110715<br />

Fig. 2. Daily raw SVI Time Series<br />

Finally, Table 1 presents the means, standard deviations and correlations of the main<br />

variables of interest. Since stock returns are measured as a percentage change and are<br />

only available for trading days, we do not include the raw measures of dowt, sentt and<br />

SVIt. Instead, we calculate the percentage change of sentiment and investor attention<br />

on trading days, report these values in the following table and include them in our<br />

further analyses. Non-trading days are excluded. As Table 1 shows, we find small<br />

positive correlations, except for SVI. In this case, SVI is negatively correlated to<br />

DJIA returns. Furthermore, we find a small negative correlation of SVI and sent.<br />

However, this correlation is not statistically significant (p = 0.84).<br />

Table 1. Means, Standard Deviations and Correlations<br />

Mean Std. Dev. dj sent sent x SVI SVI<br />

dj 0.00046 0.00073 1.0000<br />

sent 0.04286 0.02078 0.0754 1.0000<br />

sent x SVI -0.00017 0.00290 0.0657 0.0066 1.0000<br />

SVI 0.01162 0.00935 -0.3205*** -0.0118 0.1380** 1.0000<br />

* / ** / *** = significant at a 10% / 5% / 1% level of significance<br />

4.2 Impact of Media Sentiment and Investor Attention on Stock Returns<br />

To investigate the impact of media sentiment and investor attention on DJIA returns,<br />

we regress the DJIA returns (dowt) on our daily sentiment measure (sentt), investor<br />

attention (SVIt) as well as the moderating effect taking into account both variables<br />

(sentt x SVIt):<br />

Within Equation 2, εt denotes the error term. Additionally, Controlst stands for<br />

several control variables that are also included to ensure that the results are not biased<br />

because of further effects possibly influencing stock returns: To control for day<br />

patterns of stock returns and for the January effect that can cause abnormal stock<br />

returns, we include dummy variables for the different trading days as well as for<br />

20110730<br />

Day<br />

20110814<br />

20110829<br />

20110913<br />

20110928<br />

20111013<br />

20111028<br />

20111112<br />

20111127<br />

20111212<br />

20111227<br />

20120111<br />

20120126<br />

20120210<br />

20120225<br />

(2)


January [5]. Furthermore, to account for the developments within August 2011, we<br />

also include a dummy variable for this month. Additionally, we include variables for<br />

past volatility 1 , previous trading volume 2 as well as previous DJIA returns up to five<br />

lags [5]. Within the regression, we use heteroscedasticity- and autocorrelationconsistent<br />

standard errors [32]. The results of the regression are denoted in Table 2. In<br />

order to test for multicollinearity, the variance inflation factor was calculated for each<br />

independent variable. Thereby, no multicollinearity was detected since the highest<br />

score of 2.17 is below common thresholds of 4 and 10 [33].<br />

Table 2. Impact of Media Sentiment and Investor Attention on DJIA Returns<br />

Coefficient Standard Error<br />

sent 0.0033198** (0.0015118)<br />

sent x SVI 0.0323227* (0.0187595)<br />

SVI -0.0284493*** (0.0068986)<br />

* / ** / *** = significant at a 10% / 5% / 1% level of significance<br />

At first, the results confirm the impact of media sentiment on stock prices. As<br />

indicated by a positive coefficient for the sentiment measure, we can note that an<br />

increase in media sentiment leads to an increase in the corresponding DJIA return.<br />

Thereby, the coefficient is significant at a 5% level of significance.<br />

Considering the joint impact of investor attention and media sentiment on DJIA<br />

returns, we also find a positive relationship which is significant at a 10% level of<br />

significance. In this context, the positive effect of media sentiment on DJIA returns is<br />

increased when investor attention is high. Although an increased SVI does not imply<br />

that investors actually read the news articles published via DJNS, it can be noted that<br />

an augmented interest in the corresponding topic (i.e. the DJIA) prevails. Since the<br />

news articles at hand are published on several websites as well, it is more likely that<br />

the news articles are actually read by investors’ searching for information via Google.<br />

As follows, more investors are confronted with a certain level of media sentiment and<br />

consequently, their trading decisions are influenced.<br />

Interestingly, the sole impact of investor attention on DJIA returns is negative,<br />

whereas the coefficient is significant at a 1% level of significance. At a first sign, this<br />

result contradicts previous research. In this context, [22] find a positive relationship of<br />

investor attention (measured by the amount of Google searches) and stock returns.<br />

However, [22] show that, when controlling for market capitalization, the positive<br />

price pressure is only present among the smaller half of their stock sample.<br />

Furthermore, in their study, an interaction effect of market capitalization and SVI has<br />

a negative impact on returns [22]. Since the constituents of the DJIA have a high<br />

market capitalization and are analyzed on an aggregated level, these results do not<br />

contradict previous studies. Considering the control variables, the results remain<br />

robust when including a dummy variable for August 2011. Thereby, the day-of-week<br />

1 Thereby, the approach proposed by [5] is followed: to account for past volatility, the daily<br />

returns of the DJIA are demeaned to obtain a residual, this residual is squared and the past<br />

60-day moving average is subtracted.<br />

2 Specifically, the detrended logvolume is used as proposed by [31].


dummy variables as well as the dummy variable for January have no significant<br />

influence, whereas few of the lagged control variables for previous returns, volatility<br />

and trading volumes have a significant influence (not reported in Table 2 due to space<br />

constraints).<br />

5 Predicting Bidirectional Market Movements based on Media<br />

Sentiment and Investor Attention<br />

5.1 General Setup<br />

In the following, we investigate whether the influence of media sentiment and<br />

investor attention on DJIA returns can be taken into account to forecast future market<br />

movements. Thereby, we focus on predicting DJIA returns by means of machine<br />

learning techniques. In this case, machine learning techniques are advantageous<br />

because of two main reasons. At first, the evaluation becomes more reliable since<br />

evaluation methodologies like 10-fold cross validation can be used [34]. With this<br />

respect, 10-fold cross validation offers the possibility to use an increased number of<br />

data items for evaluating the trained model. Second, machine learning classifiers like<br />

Support Vector Machine (SVM) are also suitable to cover nonlinear relationships<br />

within the data which may improve forecasting results [35].<br />

For predicting DJIA returns, we make use of supervised learning and train a<br />

machine learning classifier with labeled training data in order to find patterns within<br />

the data that can serve for future predictions. Thus, every observation (i.e. each<br />

trading day) is labeled according to the corresponding DJIA return. Thereby, we<br />

assign two labels: the first label is assigned according to the contemporaneous DJIA<br />

return (T0), the second label is related to the one day ahead return (T0+1). We follow<br />

previous studies and focus on two classes [36]: the class negative is assigned if the<br />

corresponding DJIA return is lower than zero, otherwise, the class positive is<br />

assigned. In total, for T0 and T1 forecasts, 161 (131) observations are labeled as<br />

positive (negative).<br />

Within this study, we make use of a Support Vector Machine (SVM) classifier<br />

since SVMs have proven to be a good choice for financial forecasting [37]. Thereby,<br />

the same input variables have been used that were already defined in section 4.2, i.e.<br />

media sentiment, investor attention, the interaction term as well as the control<br />

variables.<br />

5.2 Evaluation<br />

To evaluate the proposed machine learning setup, we make use of 10-fold stratified<br />

cross validation [34]. In this case, the whole dataset is split into 10 subsets, whereas<br />

each subset is used k-1 times for classifier training and once for classifier testing. At<br />

the end of each iteration, a contingency table is created that contains the number of<br />

true positives (TP), false positives (FP), true negatives (TN) and false negatives (FN).<br />

Finally, a global contingency table is created by summing up the different


contingency tables (micro-averaging) [38]. Based on this global contingency table,<br />

different performance metrics are calculated. Thereby, we focus on accuracy,<br />

indicating the percentage of cases classified correctly as well as Precision, Recall and<br />

the F1-measure [39, 40]. These metrics are defined as follows:<br />

(3)<br />

(4)<br />

The results of our evaluation for the contemporaneous as well as the T0+1 forecasts<br />

are depicted in Table 3. Thereby, a SVM classifier using a radial basis function kernel<br />

(as proposed by [41]) has been made use of. The corresponding parameters have been<br />

selected via grid search 3 .<br />

Table 3. Forecasting DJIA Market Movements Using Media Sentiment and Investor Attention<br />

Class: positive Class: negative<br />

Forecast Accuracy Precision Recall F 1 Precision Recall F 1<br />

T 0 58.55 58.40 86.34 69.67 59.26 24.43 34.60<br />

T 0+1 58.56 59.71 76.40 67.03 55.81 36.64 44.24<br />

All values are given as percentages.<br />

Related to the predictability of DJIA market movements based on media sentiment<br />

and investor attention, it can be observed that the obtained results are better than<br />

results being achieved just by chance. This is evidenced by the fact that the precision<br />

scores are above 50% in all cases. Additionally, contemporaneous market movements<br />

as well as the returns of the following day can be predicted with similar accuracies of<br />

58.55% and 58.56% respectively. Considering the class recall, we find that recall of<br />

the class positive is higher than the recall of the class negative which can be attributed<br />

to the class distribution within our sample: 55% of all observations are labeled as<br />

positive and, as follows, the SVM is trained respectively.<br />

However, taking the economic value of these results into account, an accuracy of<br />

below 60% cannot be considered as promising. Thus, using only these structured<br />

variables as input data for a decision model can hardly be seen as a source for<br />

significant profits. Instead, many cases are classified incorrectly. This may be<br />

attributed to the noisy nature of financial markets and to the fact that the decision<br />

model does not take into account the textual information published within the news<br />

articles under investigation. As a consequence, investor sentiment and investor<br />

attention should not be used solely to forecast market movements. Instead, they<br />

should be incorporated in existing forecasting models to improve forecasting results.<br />

3 we followed the procedure proposed by [41] and evaluated the proposed values for C, a<br />

penalty parameter and γ, a parameter of the radial basis function. For To+1, C = 512 and γ =<br />

2 -15 lead to the best results. In the case of T 0, C= 32 and γ = 2 -15 were selected.<br />

(5)<br />

(6)


5 Summary and Conclusion<br />

In recent years, the impact of media sentiment on financial variables like stock prices<br />

has been of great interest. However, one crucial prerequisite of relating media<br />

sentiment to financial variables has not been taken into account: current studies do not<br />

consider whether the news articles expressing sentiment are actually noticed by<br />

investors. As a consequence, we examine the interplay between media sentiment and<br />

investor attention in order to investigate the joint impact of both variables on Dow<br />

Jones Industrial Average returns.<br />

Based on an empirical analysis of the sentiment expressed within 13,696 financial<br />

news articles, we find that higher investor attention increases the impact of media<br />

sentiment on DJIA returns. Thus, when investors actually pay attention to a financial<br />

instrument, the impact of media sentiment on these financial variables is higher.<br />

Furthermore, this effect is already measured at the same day rather than with a delay<br />

of several days. If the variables under investigation are used to forecast the DJIA<br />

returns, it can be observed that the results obtained are higher than results being<br />

achieved just by chance. However, there are still many cases which are classified<br />

incorrectly. As follows, further (unstructured) information has to be incorporated<br />

within the decision model in order to improve forecasting results. For example,<br />

textual inputs or technical indicators may also be considered to incorporate the<br />

information published as well as current market trends [42, 43].<br />

Within our study, media sentiment and investor attention are measured on a daily<br />

basis. As a consequence, our study does not cover intraday effects of media sentiment<br />

and investor attention on stock returns. In this context, the intraday stock price impact<br />

of media sentiment may be measured by taking into account tick-by-tick trading data.<br />

However, since SVI is only available on a daily basis, we are aware of the limitation<br />

that actually, intraday effects of both variables are not covered and should be<br />

investigated as soon as an intraday SVI is available. Additionally, within further<br />

research, the interplay between media sentiment and investor attention as well as its<br />

impact on financial variables should also be examined at a stock level. Furthermore,<br />

since the effect of investor attention among small-capitalized stocks has found to be<br />

higher [22], less frequently traded stocks will also be incorporated in our analysis.<br />

Finally, contemporary research reports that traditional news media and social media<br />

are interconnected. In this case, topics that are discussed within newspapers are also<br />

talked about within blogs [44]. Thus, the discussions within social media could also<br />

be analyzed in order to develop a more fine-grained indicator for measuring investor<br />

attention at a topic level.<br />

Acknowledgements. The research leading to these results has received funding from<br />

the European Community's Seventh Framework Programme (FP7/2007-2013) within<br />

the context of the Project FIRST, Large scale information extraction and integration<br />

infrastructure for supporting financial decision making, under grant agreement n.<br />

257928.


References<br />

1. Antweiler, W., Frank, M.Z.: Is All That Talk Just Noise? The Information Content of<br />

Internet Stock Message Boards. The Journal of <strong>Finance</strong> 59, 3, 1259–1294 (2004)<br />

2. Tetlock, P.C., Saar-Tsechansky, M., Macskassy, S.: More Than Words: Quantifying<br />

Language to Measure Firms' Fundamentals. The Journal of <strong>Finance</strong> 63, 3, 1437–1467<br />

(2008)<br />

3. Dellavigna, S., Pollet, J.M.: Investor Inattention and Friday Earnings Announcements. The<br />

Journal of <strong>Finance</strong> 64, 2, 709–749 (2009)<br />

4. Hirshleifer, D., Lim, S.S., Teoh, S.H.: <strong>Dr</strong>iven to Distraction: Extraneous Events and<br />

Underreaction to Earnings News. The Journal of <strong>Finance</strong> 64, 5, 2289–2325 (2009)<br />

5. Tetlock, P.C.: Giving Content to Investor Sentiment: The Role of Media in the Stock<br />

Market. The Journal of <strong>Finance</strong> 62, 3, 1139–1168 (2007)<br />

6. Muntermann, J., Guettler, A.: Intraday stock price effects of ad hoc disclosures: the German<br />

case. Journal of International Financial Markets, Institutions and Money 17, 1, 1–24 (2007)<br />

7. Shleifer, A., Summers, L.H.: The Noise Trader Approach to <strong>Finance</strong>. The Journal of<br />

Economic Perspectives 4, 2, 19–33 (1990)<br />

8. Black, F.: Noise. The Journal of <strong>Finance</strong> 41, 3, 529–543 (1986)<br />

9. Shleifer, A., "Inefficient markets. An introduction to behavioral finance". Oxford Univ.<br />

Press, Oxford (2000)<br />

10. de Bondt, W.F. M.: A portrait of the individual investor. European Economic Review 42, 3-<br />

5, 831–844 (1998)<br />

11. Brown, G.W., Cliff, M.T.: Investor sentiment and the near-term stock market. Journal of<br />

Empirical <strong>Finance</strong> 11, 1, 1–27 (2004)<br />

12. Shiller, R.J.: Stock Prices and Social Dynamics. Brookings Papers on Economic Activity<br />

1984, 2, 457–510 (1984)<br />

13. Loughran, T., McDonald, B.: When Is a Liability Not a Liability? Textual Analysis,<br />

Dictionaries, and 10-Ks. The Journal of <strong>Finance</strong> 66, 1, 35–65 (2011)<br />

14. Das, S.R., Chen, M.Y.: Yahoo! for Amazon: Sentiment Extraction from Small Talk on the<br />

Web. Management Science 53, 9, 1375–1388 (2007)<br />

15. Bollen, J., Mao, H., Zeng, X.-J.: Twitter mood predicts the stock market. Journal of<br />

Computational Science 2, 1, 1–8 (2011)<br />

16. Barber, B.M., Odean, T.: All That Glitters: The Effect of Attention and News on the<br />

Buying Behavior of Individual and Institutional Investors. Review of Financial Studies 21,<br />

2, 785–818 (2008)<br />

17. Busse, J.A., Green, T.C.: Market efficiency in real time. Journal of Financial Economics<br />

65, 3, 415–437 (2002)<br />

18. Lou, D.: Attracting investor attention through advertising. Discussion paper 644. Financial<br />

Markets Group, London School of Economics and Political Science, London, UK. (2009)<br />

19. Chemmanur, T.J., Yan, A.: Advertising, Attention, and Stock Returns. Working Paper.<br />

Available at SSRN: http://ssrn.com/abstract=1340605 (2009)<br />

20. Cohen, L., Frazzini, A.: Economic Links and Predictable Returns. The Journal of <strong>Finance</strong><br />

63, 4, 1977–2011 (2008)<br />

21. Louis, H., Sun, A.: Investor Inattention and the Market Reaction to Merger<br />

Announcements. Management Science 56, 10, 1781–1793 (2010)<br />

22. Da, Z., Engelberg, J., Gao, P.: In Search of Attention. The Journal of <strong>Finance</strong> 66, 5, 1461–<br />

1499 (2011)<br />

23. Pang, B., Lee, L., Vaithyanathan, S.: Thumbs up? Sentiment Classification using Machine<br />

Learning Techniques. Proceedings of the Conference on Empirical Methods in Natural<br />

Language Processing, Philadelphia, Pennsylvania, USA, 79–86 (2002)<br />

24. Pang, B., Lee, L.: Opinion Mining and Sentiment Analysis. Foundations and Trends in<br />

Information Retrieval 2, 1-2, 1–135 (2008)


25. Zhou, L., Chaovalit, P.: Ontology-supported polarity mining. Journal of the American<br />

Society for Information Science and Technology 59, 1, 98–110 (2008)<br />

26. Thelwall, M., Buckley, K., Paltoglou, G.: Sentiment in Twitter events. Journal of the<br />

American Society for Information Science and Technology 62, 2, 406–418 (2011)<br />

27. Zhang, W., Skiena, S.: Trading Strategies To Exploit Blog and News Sentiment.<br />

Proceedings of the 4th International AAAI Conference on Weblogs and Social Media,<br />

Washington, DC, USA (2010)<br />

28. Ginsberg, J., Mohebbi, M.H., Patel, R.S., Brammer, L., Smolinski, M.S., Brilliant, L.:<br />

Detecting influenza epidemics using search engine query data. Nature 457, 7232, 1012–<br />

1014 (2009)<br />

29. Penman, S.H.: The distribution of earnings news over time and seasonalities in aggregate<br />

stock returns. Journal of Financial Economics 18, 2, 199–228 (1987)<br />

30. Patell, J.M., Wolfson, M.A.: The intraday speed of adjustment of stock prices to earnings<br />

and dividend announcements. Journal of Financial Economics 13, 2, 223–252 (1984)<br />

31. Campbell, J.Y., Grossman, S.J., Wang, J.: Trading Volume and Serial Correlation in Stock<br />

Returns. The Quarterly Journal of Economics 108, 4, 905–39 (1993)<br />

32. Newey, W.K., West, K.D.: A Simple, Positive Semi-Definite, Heteroskedasticity and<br />

Autocorrelation Consistent Covariance Matrix. Econometrica 55, 3, 703–708 (1987)<br />

33. O'Brien, R.M.: A Caution Regarding Rules of Thumb for Variance Inflation Factors.<br />

Quality & Quantity 41, 673–690 (2007)<br />

34. Mitchell, T., "Machine learning". McGraw-Hill, London (1997)<br />

35. Dhar, V., Stein, R., "Intelligent decision support methods. The science of knowledge<br />

work". Prentice Hall, Upper Saddle River, NJ (1997)<br />

36. Groth, S.S., Muntermann, J.: Supporting Investment Management Processes with Machine<br />

Learning Techniques. Proceedings of the 9th Internationale Tagung Wirtschaftsinformatik,<br />

vol. 2, Vienna, Austria, 275–284 (2009)<br />

37. Tay, F.E. H., Cao, L.: Application of support vector machines in financial time series<br />

forecasting. Omega 29, 4, 309–317 (2001)<br />

38. Chau, M., Chen, H.: A machine learning approach to web page filtering using content and<br />

structure analysis. Decision Support Systems 44, 2, 482–494 (2008)<br />

39. Hotho, A., Nürnberger, A., Paaß, G.: A Brief Survey of Text Mining. GLDV Journal for<br />

Computational Linguistics 20, 1, 19–62 (2005)<br />

40. van Rijsbergen, C.J., "Information retrieval", 2nd ed. Butterworths, London (1979)<br />

41. Hsu, C.W., Chang, C.C., Lin, C.J.: A practical guide to support vector classification.<br />

National Taiwan University, http://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf<br />

(2011-10-16) (2003)<br />

42. Mittermayer, M.-A.: Forecasting Intraday Stock Price Trends with Text Mining<br />

Techniques. Proceedings of the 37th Hawaii International Conference on System Sciences,<br />

Big Island, Hawaii, USA (2004)<br />

43. Lam, M.: Neural network techniques for financial performance prediction: integrating<br />

fundamental and technical analysis: Data mining for financial decision making. Decision<br />

Support Systems 37, 4, 567–581 (2004)<br />

44. Leskovec, J., Backstrom, L., Kleinberg, J.: Meme-tracking and the Dynamics of the News<br />

Cycle. Proceedings of the 15th International Conference on Knowledge Discovery and Data<br />

Mining (2009)


MEASURING THE INFLUENCE OF PROJECT<br />

CHARACTERISTICS ON OPTIMAL SOFTWARE PROJECT<br />

GRANULARITY<br />

Weber, Moritz Christian, Goethe University Frankfurt, Campus Westend, Grüneburgplatz 1,<br />

60329 Frankfurt, Germany, moweber@wiwi.uni-frankfurt.de<br />

Wondrak, Carola, Goethe University Frankfurt, Campus Westend, Grüneburgplatz 1,<br />

60329 Frankfurt, Germany, cwondrak@wiwi.uni-frankfurt.de<br />

Abstract<br />

With a well-structured design of reusable software, it is possible to save costs in future software<br />

development cycles and generate added value. Nowadays it is difficult to quantify which grade of<br />

project structuring enhances its reusability. This is the question this paper aims to examine, based on<br />

the financial Portfolio Selection theory and Value at Risk methods that can be adapted to the field of<br />

software project structuring. Using these methods, the value of the reusable source code can be<br />

evaluated and the risks of reimplementation be compared with alternative code structures. So we can<br />

quantify and measure the effects of the software project structure on the risks of reimplementation.<br />

This model is applied to 27 Github open source software projects and enables us to investigate the<br />

influences of the project characteristics on the best possible software project granularity.<br />

Keywords: Project Management, IT Project Portfolio, Software Granularity, Value at Risk.


1 Motivation<br />

Static and monolithic information systems mark the beginning of enterprise IT infrastructures (Krafzig<br />

et al., 2004). With the emergence of spreading network structures, these inflexible infrastructures are<br />

split into subsystems and managed as interrelated parts of an IT portfolio (Erl, 2005). ‘The foundation<br />

of an IT portfolio is the firm’s information technology’ (Weill and Vitale, 2002). This is — among IT<br />

components and human infrastructures — defined by shared IT services and standard applications<br />

(Weill and Vitale, 2002). Not only software development, but also general strategy research reflects<br />

synergies and diversifications of portfolios. These allow the generation of additional positive<br />

economic values gained by the merging and restructuring of financial portfolios (Amihud and Lev,<br />

1981; Christensen and Montgomery, 1981; Farjoun, 1998). As the realization of synergistic economies<br />

is crucial, it can be seen as a third pillar next to financial and vertical economics (Hill and Hoskisson,<br />

1987). Markowitz’s Portfolio Selection theory (1952) defines a theoretical framework for portfolio<br />

diversification and synergies which can be extended (Myers 1982) towards strategic planning and has<br />

the potential to be applied in other research domains (Rockafellar and Uryasev, 1999). As IT projects<br />

are also aggregated in portfolios, this theory supports evidence as to how software portfolios should be<br />

structured.<br />

Synergies typically depend on the size and structure of the portfolio (Eckbo, 1983; Bradley et al.,<br />

1983). Two kinds of synergies can be distinguished for IT portfolios (Cho and Shaw, 2009;<br />

Tanriverdi, 2005; 2006): On the one hand bundling two components together allows sub-additive cost<br />

savings within similar processes (Teece, 1980). On the other hand, two interrelated components can<br />

actually enable an additional, super-additive benefit at the same time as increasing the effort for both<br />

(Milgrom and Roberts, 1995). Economies of scope enhance the potential of those synergies (Teece,<br />

1980; 1982) and the ability to diversify a portfolio (Willig, 1978). This gives reason to expect lower<br />

risk costs and higher generated added value (Panzar and Willig, 1981), also for IT project portfolios.<br />

Particularly the risk stemming from size and the structure of a portfolio can be expressed in<br />

granularity. In the following sections granularity measurement is defined as ’quantifying the<br />

contribution of [...] concentrations to portfolio risk’ (Lütkebohmert and Gordy, 2007). Unlike the<br />

monolithic enterprise IT infrastructures at the beginning, it has not yet been investigated how granular<br />

and diversified a software portfolio should be from an economical point of view (Friedl, 2011). It is<br />

said that two thirds of all software projects fail or are not finished within the fixed requirements or the<br />

given time. This is also due to organizational failures in software development (Standish, 2011). That<br />

is why we ask the following research questions:<br />

� How granular should software project structures be?<br />

� How can project-granularity levels be measured and compared?<br />

� How can an optimum level of granularity be determined at present, so that good reusability<br />

and lower change risks can be expected?<br />

� Which project characteristics have a critical influence on the optimum of granularity?<br />

In the following section, we reflect the Portfolio Selection theory and the Value at Risk method (2.1)<br />

as theoretical framework to value portfolio risks. Related work on IT portfolio risk, project<br />

management (2.2) and granularity analysis (2.3) let us derive research hypotheses on risk costs, project<br />

characteristics and granularity from literature. To analyze these hypotheses, we adopted Portfolio<br />

Selection and Value at Risk to measure quantitatively risk costs in software project data (3.1 and 3.2).<br />

Alternative levels of project granularities are simulated by three methods (3.3) that are applied to<br />

software version control data for the 27 most observed Github (2011) open source projects. The Value<br />

at Risk is measured for each method and each simulated project granularity level. Accordingly, a<br />

regression setup is developed to explain the influences of project characteristics on optimal software<br />

project granularity (3.4). As a result alternative project granularities are selected to show and discuss


corresponding changes of risk costs (4.1, 4.2 and 4.3). Changes in the project characteristics are<br />

measured over time for all projects and each version. These changes are taken to explain their<br />

influence to the granularity level and the minimal risk costs of software projects (4.4). Finally we<br />

summarize our results, highlight limitations and provide an outlook on future research (5).<br />

2 Related work<br />

As it is often claimed that software projects are structured like portfolios with component assets, we<br />

align our methodology with the Portfolio Selection theory (Markowitz, 1952) and the risk<br />

measurement approach called Value at Risk (Wolke, 2008) in Section 2.1. This economic perspective<br />

is applied to IT project portfolio management and risk management to link the theory to the scope of<br />

information systems research (Section 2.2). It is often assumed that software development projects –<br />

as a subset of IT projects – have a high risk of failure (Standish, 2011). That is why we supplement our<br />

review of related work by recent studies on structuring of software development projects and<br />

analyzing the quality of software granularity (Section 2.3).<br />

2.1 Portfolio Selection and Value at Risk<br />

A software project contains several components that are linked with each other. This is similar to a<br />

financial portfolio where each portfolio component generates a value contribution. Since software can<br />

be a substantial long-term investment, it is important to select portfolio components in the most<br />

efficient way, so that a high level of reusability and lower maintenance costs can be anticipated.<br />

Credit portfolio<br />

short<br />

fall<br />

short<br />

fall<br />

short<br />

fall<br />

short<br />

fall<br />

likely<br />

VaR<br />

95%<br />

short<br />

fall<br />

likely<br />

VaR<br />

95%<br />

short<br />

fall<br />

likely<br />

VaR<br />

95%<br />

risky positive very<br />

positive<br />

Changes<br />

risky positive very<br />

positive<br />

Changes<br />

risky positive very<br />

positive<br />

Changes<br />

How many credits of which size are risk optimal?<br />

Select the portfolio<br />

granularity with the<br />

lowest shortfall risk<br />

Figure 1. Portfolio diversification and risk in the context of correlation effects<br />

Granularity<br />

optimal<br />

Value at Risk<br />

A theoretic-based approach for the selection of components of a financial portfolio is the Portfolio<br />

Selection theory developed by Harry Markowitz (1952). It outlines how to structure an optimal<br />

portfolio with minimal risks by efficiently dividing it into partial investments (so called<br />

diversification). Thereby, interrelated influences between the components are taken into account and<br />

assembled in a single structure, so that the total risk of value depreciation is minimized (so called<br />

correlation). The value depreciation can be estimated using the Value at Risk (VaR) approach<br />

(Albrecht and Maurer, 2005) (see Figure 1). It ‘can be applied to any risk-factor model of portfolio [...]<br />

risk, and [is ...] a very general framework’ (Lütkebohmert and Gordy, 2007). Hereby, the maximum<br />

value depreciation is a value that will not be exceeded in a given time period with a given probability<br />

(Wilkens, 2001; Wolke, 2008). The VaR approach calculates not only the expected value of losses<br />

(Expected Loss), but also explicitly outlines the unexpected losses (Campbell et al., 2001). Like this, it<br />

enables us to include the extreme cases of value depreciation in our estimations.<br />

This concept can be applied to any single evaluation, but also to multiple investment assets and entire<br />

investment portfolios. If there are multiple investments with considerable risks in one portfolio,<br />

additional synergy effects need to be taken into account (Wiedemann, 2002). Value fluctuations of<br />

these components could be both chances and risks for the whole portfolio. These risks can be


diversified by combining components which compensate the interrelated value depreciation risks. The<br />

goal is to find synergies that compensate the risks among themselves (Rockafellar and Uryasev, 1999).<br />

Through systematic validation of different portfolio structures, it is possible to assess and valuate the<br />

risk of value depreciation using the VaR approach. In the case of inefficient combinations, the so<br />

called ‘clustering risks’ can emerge (Huschens and Stahl, 2004). This kind of insufficient<br />

granularisation leads to the danger of a higher, diversified VaR (Albrecht and Maurer, 2005). The<br />

Portfolio Selection theory allows evaluating the efficient selection of portfolio components, so that the<br />

risk costs of value depreciation will decline for the future period. Rockafellar and Uryasev (1999)<br />

explicitly suggest that the VaR concept should be applied as methodology in other areas. Outside of<br />

the financial research domain we adopt this for managing the portfolio risk of IT projects.<br />

2.2 IT portfolio risk and project management<br />

Cho and Shaw (2009) model and simulate the influences of synergies on the IT portfolio selection by<br />

applying the Markowitz theorem and the VaR approach to the research scope of IT portfolios. They<br />

find that not only portfolio returns, but also portfolio risks need to be taken into account. Moreover,<br />

they argue that only few studies investigated the influences of IT synergy on portfolio risks. Benaroch<br />

(2002; 2006) analyzes IT risks and their influence on large portfolios of IT investments. Strong<br />

evidence is found that IT risk management is largely driven by intuitions, which leads to suboptimal<br />

and counterproductive results. Tanriverdi and Ruefli (2004) state that synergies of IT projects also<br />

influence the risk of those. A failure of one IT portfolio component may cause failures in other parts of<br />

the portfolio, which may increase the risks at higher levels of diversification (Tanriverdi, 2005; 2006).<br />

This indicates that the following hypothesis on software project granularity should be rejectable:<br />

Hypothesis H10: Risk costs have no crucial influence on optimal software project granularity.<br />

The organization of software development project portfolios can be improved by a well coordinated<br />

project management. Lichter and Mandl-Striegnitz (1999) show that most software development<br />

projects fail because of inappropriate organizational settings and not because of the technical<br />

implementations. In addition, inefficient processes may cause design errors and an unexpected<br />

increase in project costs. A rule of thumb on how the cost of a single present design error influences<br />

future development and refactoring expenditures is given by so called ‘Rule of Ten’ (Pfeifer, 2001). It<br />

estimates that an undiscovered design error multiplies the required cost of rectification by ten for each<br />

following period. It is possible to recognize design errors earlier by using efficient project<br />

management and verifiable key figures (Lichter and Mandl-Striegnitz, 1999).<br />

Especially the reuse of previously implemented software artefacts is an opportunity to reduce costs<br />

and work effort in multiperiodic software projects. Necessarily the design of software artefacts should<br />

not be too complex and problems with complex interdependence between the artefacts should be taken<br />

into account (Schirmer and Zimmermann, 2008). This involves both, dependencies between different<br />

parts of the software and a good kind of granularity. More specific, special attention must be paid to<br />

project characteristics, the structure of the portfolio and the selection of subprojects (Schirmer and<br />

Zimmermann, 2008). So we expect to reject the following null hypothesis:<br />

Hypothesis H20: Project characteristics have no influence on optimal software project granularity.<br />

2.3 Granularity analysis<br />

Different approaches analyze optimal software project structures (Krafzig et al., 2004, p.18). Thereby,<br />

efforts are being made to quantify the quality of structures by measuring the utility and monetary value<br />

to software structuring decisions: Erradi et al. (2006) propose a quantitative ranking with numerical<br />

steps up to 10, which considers the level of dependencies between the project components.<br />

Furthermore, they criticize the lack of theory that would support the findings of correct granularity for<br />

maximal component reuse and lower project risks (Erradi et al., 2006). Alternative research strategies<br />

evaluate the outsourcing and awareness potential of distributed projects and identify project


structuring candidates by an economical analysis (Klose et al., 2007). In their empirical analysis<br />

Braunwarth and Friedl (2010) took the internal cost allocation of a company as a benchmark and<br />

simulate potential structures of granularity. In addition of this study, Friedl classified granularity<br />

approaches based on a literature review and gave advice for designs of good granularity on typical<br />

factors of influence in software development projects (Friedl, 2011). Katzmarik supports the<br />

granularity decision with a micro-economical approach, in which the project structures are modeled as<br />

‘Software as a Service’-products (SaaS) and evaluated (Katzmarzik, 2011). Therefore, we assume that<br />

the following null hypothesis cannot be supported:<br />

Hypothesis H30: Optimal software IT project granularity cannot be estimated empirically.<br />

3 Methodology<br />

In the following section the VaR concept is applied to software development projects by using version<br />

change data from software control systems. The software project structures as well as the historical<br />

variance and covariance of source code lengths are measured. Through three different methods<br />

alternative granularity structures are simulated and the level of minimal change expenditure risk is<br />

determined by VaR valuation. Finally, this minimal level is regressed against the given project<br />

characteristics to measure the influences of the project parameters on the optimal level of granularity.<br />

3.1 Assumptions<br />

In accordance with the Portfolio Selection theory (Markowitz, 1952) it is assumed that each portfolio<br />

can be diversified by a requirement-optimal or risk-minimal allocation, if a value can be determined.<br />

Modern software engineering approaches separate classes, services and other entities into single files<br />

(Erl, 2005). Therefore, we assume that each change of a file is an implicit change of a portfolio<br />

component. Agile software development practitioners (like in open source projects) claim that each<br />

change of a software project implements a single functional requirement or bug fix, which is<br />

documented in our case by the version control system. For simplification, it is assumed that the change<br />

expenditure of a file is highly correlated with the number of changed and unchanged lines of code<br />

(LOC). It is further assumed that the kind of development and refactoring is anchored in the project<br />

characteristics and that it stays similar in the history of a project. That is why historical change data is<br />

a good estimate of future change risks. The current size of a source code file is taken as present value.<br />

Using historical recorded project development costs, the LOC calculation can easily be transformed<br />

into a utility measure by applying COCOMO methodology (Katzmarzik, 2011, p. 23).<br />

3.2 Data aggregation and measuring method<br />

The data of software version control systems is used to measure direct development structures and<br />

characteristics of software projects. Thereby, the historical sizes of each file as well as realized<br />

software changes are measured. So the extent of each requirement implementation or bug fix can be<br />

precisely determined. Each requirement implementation can affect multiple file changes. Like Grigore<br />

and Rosenkranz (2011), Turek et al. (2010) and Wierzbicki et al. (2010) we measure the residual value<br />

of reusable source code: This is the number of document lines which are created by one author and<br />

remain in the document after the document revision (adds/removes). This residual value of reusable<br />

code is identified for each file of every requirement implementation (patch). This value is saved with<br />

the residual values of the corresponding files of each change (see Figure 2 ‘Data aggregation’).<br />

The VaR concept is applied (Wiedemann, 2002) to identify at which structural level the residual value<br />

of reusable source code remains the highest possible. This determines the loss that is not exceeded in<br />

the covered period of time with a given probability. Thus the calculation also takes historical<br />

fluctuations and correlative dependencies between the components of a portfolio into account. Here,<br />

the fluctuations of the residual value of reusable code (determined in the data aggregation) are used.<br />

For all pairwise changed files, a covariance value is calculated and aggregated in a covariance matrix


(Wiedemann, 2002). The risk value is calculated by matrix multiplications of the file lengths at the<br />

current point in time and the variance-covariance matrix of relative residual values of reusable source<br />

code. Finally, the risk quantile for the given probability is determined (Wilkens, 2001; Wolke, 2008)<br />

(see Figure 2 ’Measuring method’).<br />

Version control system<br />

Data aggregation Measuring method Simulation<br />

c<br />

a<br />

b<br />

Residual value of reusable source<br />

code<br />

=<br />

Lines per file – all changes<br />

t0 t1 t2 t3<br />

b<br />

a<br />

Original<br />

file<br />

Covariance of value<br />

fluctuations<br />

t0 t1 t2 t3<br />

Covariance<br />

Reusable source code<br />

- +<br />

New version<br />

after revison<br />

divVaR = LOC*cov(a,b,..)*(LOC)’<br />

Maximum loss of residual value with<br />

given probability in a period<br />

Original corr.<br />

(tar method)<br />

a b c d<br />

1<br />

10 -1<br />

10 -2<br />

.<br />

.<br />

e f g h<br />

Strong corr.<br />

(corr method)<br />

a<br />

e<br />

b<br />

d<br />

Increasing granularity<br />

Search for optimal granularity by<br />

valuing the minimal diversified<br />

Value at Risk<br />

c<br />

Weak corr.<br />

(rand methode)<br />

a,c,g,h,..<br />

1 2 3 4 5 6 7 8<br />

The worst granularity level is normalized to 1 for validation of<br />

the ten spot rule<br />

Figure 2. Data aggregation, measuring method and simulation of the analysis<br />

3.3 Simulation<br />

The granularity of every historic grown project is given implicitly by the project structure. Therefore,<br />

a diversified VaR can be calculated directly. It quantifies risk of reimplementation for the given<br />

granularity structure. To evaluate the risk of alternative granularity levels, variants with comparable<br />

structures must be derived. A sufficient effective method (hereinafter ‘tar method’) starts with the<br />

lowest folder level and aggregates all files per folder into a single file. Afterwards this file is moved to<br />

the superior folder (Lee et al., 2007). This way, granularity variants can be generated which keep the<br />

same basic structure. In a simplification of Lee’s approach, the common tar algorithm is recursively<br />

applied (FSF, 2006). With source code packages and functional structuring of folders the dependency<br />

structures and conditional changes remain unaltered even for extensive aggregations. The most coarse<br />

and presumable worst granularity generated by this method, is to copy all source code into a single<br />

file. For each aggregation level, the size of files and changes as well as the covariance are remeasured<br />

and a risk value is calculated using the VaR approach.<br />

The results of tar methodology are validated by methodological triangulation using two additional<br />

control methods. The first validation method (‘corr method’) selects an initial file from the project for<br />

each simulation run (Blobel and Lohrmann, 1998). Starting with this file, dependently changed files<br />

are recursively added. This is repeated until the number of files fits the amount of files of the<br />

corresponding granularity level of the tar method. The second method (‘rand method’) selects random<br />

files, but does not consider correlation structures. Thus, the first control method simulates subprojects<br />

of the entire project with a strong correlation. The second method simulates the omission of these<br />

correlations. Huschens and Stahl (2004) show that changes in the correlation structures influence the<br />

relative evaluation of granularities only slightly. The structure of the entire project is approximated by<br />

averaging the partial results (Blobel and Lohrmann, 1998).<br />

The validation methods yield good comparative values. But it should be kept in mind that these<br />

methods are too aggressive and too coagulative in their simulation/estimation, due to their random<br />

selection and partial simulation. As a result, badly diversified and extreme project structures can be<br />

generated. This is why these two validation methods are used for triangulation reasons only and not<br />

for measuring the absolute risk levels of the relative residual value of the reusable source code. To<br />

facilitate comparison — also with the ‘Rule of Ten’ (Pfeifer, 2001) — we start at the granularity with<br />

Risk<br />

costs<br />

Granularity


the highest VaR for each method. Subsequent all VaRs are normalized to this initial value of 1. The<br />

normalized value of 1 is usually the case, if just one file contains all source code (Katzmarzik, 2011,<br />

p.29) (see Figure 2 ‘Simulation’).<br />

3.4 Regression setup<br />

To investigate our research hypotheses and the influences of given project characteristics on the<br />

optimal project structure, we regress typical project key figures on the optimal project granularity of<br />

each release version (see Equation 1). As the time and the amount of version releases is not<br />

synchronized between the 27 projects, we build a cross-sectional dataset containing the interversion<br />

differences in optimal granularity (OptGranu) and the risk costs at optimal granularity (RiskCosts).<br />

Additionally typical project characteristics like size units (Katzmarzik, 2011, p.8) and time measures<br />

(Pfeifer, 2001) are added. These are number of patches (Patches), average file changes per patch<br />

(AvgPFiles), number of files (Files), total lines of code (LOC) and duration of the version release<br />

(Time) are added. To harmonize sizes of increasing and decreasing changes, all differences are<br />

logarithmized. Moreover, we added control variables (Controls) for each project to identify project<br />

specific influences.<br />

The sample is tested for a typical structural bias. Considering the assumably close dependency of the<br />

exogenous variables, no multicollinearity is observed. Since heteroscedasticity of the residuals<br />

occurred in the original model, robust standard errors are applied.<br />

4 Results<br />

4.1 Descriptive results<br />

Due to easy accessibility and publishability our data sample consists of top ranked Github open source<br />

projects. However, this setup can also be easily applied for commercial projects like SOA, SaaS or<br />

proprietary software. Github ranks the interest in projects by counting the number of observers<br />

following the project and the number of projects that are build upon this project. We selected 27<br />

projects and ensured that the sample consists of equally distributed programming paradigm styles<br />

(scripting, imperative/functional and object-oriented), execution styles (native, byte code, interpreted)<br />

and software category (webserver/ -framework, database/ libraries, system/ languages). The<br />

occurrences of each language, the type and paradigms of projects within this sample can be seen in<br />

Table 1:<br />

Exec native bytecode interpet Lang c c++ java objc erlang python ruby js<br />

27 9 9 9 27 6 3 3 3 3 2 2 5<br />

Categ web db/lib sys/lan Type sys lang db lib webserver webframe<br />

27 9 9 9 27 6 3 7 2 2 7<br />

Style oo script im/fun Parad oo script imp func<br />

27 9 9 9 27 9 9 6 3<br />

Table 1. Descriptive statistics of the selected sample of projects<br />

Overall, we extracted 1625 observations (version tags). Some projects like ‘postgres’ have a long<br />

version history, others like ‘nu’ displayed a history of continuous integration into a single master<br />

version. As we want to observe the influence of project characteristics on the optimal project<br />

granularity, we accept also these extreme project settings. Project-specific influences are tested with<br />

the aid of control variables in the final regression. Descriptive figures are listed in Table 2 and 3.<br />

(1)


Name<br />

CassandraClojure<br />

Couch<br />

DB<br />

Django<br />

Ejabberd<br />

Git<br />

Gitx<br />

Httpd<br />

jQuery<br />

Memcached<br />

Mongo<br />

DB<br />

Node<br />

JS<br />

Language java java erl python erl c objc c js c c++ c++ c erl<br />

Type db lang db webfr sys sys sys webse webfr db db sys lang lib<br />

Paradigm oo oo func script func imp oo imp script imp oo oo imp func<br />

Versions 54 3 5 23 28 296 11 174 63 30 104 105 1 6<br />

Fold. Avg. 204 39 70 3152 36 178 87 187 32 23 83 359 103 2109<br />

Files Avg. 870 250 512 4087 336 2242 330 2905 155 135 1703 4968 379 9754<br />

Table 2. Overview of project characteristics (part1)<br />

Due to page limitations and a better understanding of the calculation results, we highlight the findings<br />

for the popular web framework (Ruby on) Rails as an example. It shows average figures in its project<br />

statistics and the calculation yields similar patterns and results for the rest of 26 projects. In the final<br />

regression all 1625 observations are taken as single events. However, for clarity we aggregate all the<br />

change data of the 116 versions in Rails and calculated the all-time VaR for each granularity level.<br />

Name Phone-<br />

gap<br />

Postgres <br />

Prototype<br />

Rails<br />

Redis<br />

Ruby<br />

Scriptaculous<br />

Solr<br />

Sparkle<br />

Three<br />

20<br />

Tornado<br />

Language js c js ruby c ruby js java objc objc python c++ js<br />

Type webfr db webfr webfr db lang webfr db sys lib webse sys webfr<br />

Paradigm script imp script script imp script script oo oo oo script oo script<br />

Versions 5 260 11 116 61 73 3 6 4 6 9 157 22<br />

Folders Avg. 76 380 47 777 45 459 25 1392 49 205 57 56 1587<br />

Files Avg. 113 3826 130 2055 275 3460 111 4647 297 1159 185 1646 4358<br />

Table 3. Overview of project characteristics (part2)<br />

Figure 3 illustrates that starting with a normalized VaR of 1, the VaR of all three methods declines<br />

with an increase in granularity levels. As expected, the control methods (rand and corr) calculate<br />

structures with higher VaRs. The tar method shows a similar slope to the ‘Rule of Ten’, but at the<br />

granularity level of 5 the VaRs stabilize at a stationary level. All three methods indicate the optimal<br />

VaR at a granularity level of 7. Even though the ‘Rule of Ten’ does not depend on the project<br />

characteristics of Rails, it approximates the tar graph to the granularity of 5 in the stationary region.<br />

Consequently, the control methods cannot predict correct risk cost but they estimate the same optimal<br />

granularity level as the tar method. The ‘Rule of Ten’ cannot predict the level of granularity correctly,<br />

yet it indicates a good benchmark for the risk costs predicted by the VaR.<br />

4.2 Rule of Ten as cost measure for incorrect granularity<br />

In general, the slope of the graphs shows that the theoretical ‘Rule of Ten’ underestimates the risk<br />

costs, with increasing levels of granularity. Additionally, it can be easily observed that the ‘Rule of<br />

Ten’ approximates the risk cost value of the tar method best towards the optimum. Simultaneously, the<br />

project shows that this rule of thumb noticeably underestimates the risk costs between the levels of the<br />

worst and optimal granularity. This is especially the case when multiple levels of granularity have<br />

measurable similar risk costs. Moreover, it can be argued that the validation methods simulate more<br />

aggressively, coarsely and result in worse diversified portfolios than the ‘Rule of Ten’ and the tar<br />

method. It is in line with the results of other empirical studies that even the randomized validation<br />

methods confirm the minimum risk costs found by the tar method (Huschens and Stahl, 2004).<br />

Comparing with the ‘Rule of Ten’, it can be argued that the tar method estimated the risk costs best,<br />

Nu<br />

V8<br />

OTP<br />

YUI3


ut also the validation methods determine the same level of granularity as optimal, when the absolute<br />

risk cost values are negligible.<br />

Figure 3. Analysis of Rails project results<br />

4.3 Marginal utility of granularity<br />

The data gives insights into the risk cost decline with an increase in granularity and subsequent<br />

stabilization at a stationary level. It can be intuitively understood that projects are structured with best<br />

practice methodologies and iterative approaches (Erl, 2005) until a manageable software foundation is<br />

formed. Coevally, it can be seen at the stationary levels of the project that more or less adequate level<br />

of granularity can already be discovered at lower levels of granularity. The marginal utility of<br />

additional granularity towards the optimum is declining due to complexity of coordination between the<br />

components. This can be explained by the effect of overspecification which is mentioned in literature<br />

(Millard et al., 2009). If software structures are initially designed for good reusability, but then never<br />

reused, then they might result in an inefficient overspecification and overhead. The measurement of all<br />

projects shows, that a slight over- and underspecification can be tolerable. On the one hand the risk<br />

cost increase slightly around the area of optimal granularity, while this increase is much smaller then<br />

at lower or even monolithical granularity.<br />

4.4 Influence of project characteristics on optimal granularity<br />

Figure 3 illustrates that the implicit ‘Rule of Ten’ assumption of a log-log dependency of risk costs<br />

and granularity is too unprecise. That is why we regress all 1625 version observations of the 27<br />

projects as a cross-sectional dataset of first differences of log and investigate influences of additional<br />

project characteristics. Results of the regression analysis explain changes of risk costs at optimal<br />

granularity. The number of patches, files per patch and total number of files have a highly significant<br />

influence on the change of optimal granularity. Even the influence of LOC changes is significant at<br />

least at a 5% level and time at a 10% level of significance. While risk costs, patches, files and LOC<br />

changes have a positive impact on the changes in level of optimal granularity, changes of files per<br />

patch and time have a negative influence. As all project-specific control variables do not show any<br />

significance, they are removed from the regression equation and the final results (see Table 4).<br />

All measured project characteristics are significant and allow us to reject null hypothesis H20 that there<br />

are no influences by the project characteristics. As the impact of chances in risk costs shows the<br />

highest t-value null hypothesis H10 is rejected as well. The whole regression model has an adjusted Rsquared<br />

value of 54.05% for the analyzed cross-sectional dataset. Appling an F-test for the significant<br />

influence of all coefficients of the model yields an F-Value of 273.62, which is highly significant for<br />

1625 observations and enables us to reject also null hypothesis H30. From a project management point<br />

of view, this gives evidence that increases in project parameters like project risk costs, number of<br />

patches, files and total lines of code cause a need of more granular structuring. Risk cost influence has<br />

the highest significance of the measured project characteristics. If the number of files per patch<br />

increases in a current version, it indicates too granular project structures and results in a need for a less


granular structuring. A shortage of release cycles and version snapshots shows similar results: if the<br />

time between two releases decreases, the need for more granular and more diversified project<br />

structures increases.<br />

Const Risk Costs Patches Files/Patch Files LOC Time<br />

Coefficient -0.0118 0.0229 0.0885 -0.3272 1.0451 0.1976 -6.5603<br />

Std. Error 0.0135 0.0011 0.0083 0.0397 0.3231 0.0775 3.9047<br />

t-Value -0.8728 20.8817 10.6532 -8.2349 3.2348 2.5486 -1.6801<br />

P-Value 0.3829


from project source code data. The deducted model has an explanatory power of 54%, which means<br />

46% of the observed influences are caused by influence that cannot identified with the facts from data<br />

set. The data set contains only open source projects. The measurement methods are just able to<br />

simulate the granularity between 1 and the given project granularity. Even with the deducted model an<br />

initial project version snapshot is needed to determine a good level of granularisation. In further<br />

research, we will address these limitations and deepen our analysis of the relation between risk costs<br />

and optimal granularity as these have the highest significance in our regression model. Depending on<br />

data accessibility and publishability we plan to shift our data scope from open source to SOA, SaaS<br />

and proprietary software projects to support structuring decisions in commercial software projects. We<br />

also want to investigate the dynamics of granularity over time to understand how previous versions<br />

influence the current optimal granularity.<br />

Finally, concerning the research questions, it is has been shown that the optimal granularity structure<br />

of software projects, can be measured at present. A method has been presented that enables us to<br />

measure and compare the quality of software project granularities and the influence on the optimal<br />

level of granularity. Consequently, it becomes possible to estimate quantitatively how granular<br />

software projects should be structured and how this granularity must be adjusted when specific project<br />

parameters change. We thankfully acknowledge the support of Martin Haferkorn and the E-<strong>Finance</strong><br />

<strong>Lab</strong>, Frankfurt for this work.<br />

Literature<br />

Albrecht, P. and Maurer, R. (2005). Investment- und Risikomanagement: Modelle, Methoden,<br />

Anwendungen. Schäffer-Poeschel, 2. Edition.<br />

Amihud, Y. and Lev, B. (1981). Risk reduction as managerial motive for conglomerate mergers. Bell<br />

Journal of Economics (12), pp. 605-617.<br />

Benaroch, M.(2002). Managing information technology investment risk: A real options perspective.<br />

Journal of management information systems, (19:2), pp. 43–84.<br />

Benaroch, M., Lichtenstein, Y. and Robinson, K. (2006). Real options in information technology risk<br />

management: an empirical validation of risk-option relationships. MISQ, (30:4), pp. 827 – 864.<br />

Blobel, V. and Lohrmann, E. (1998). Statistische und num. Methoden der Datenanalyse; Teubner.<br />

Bradley, M., Desai, A. and Kim, E.H. (1983). The rationale behind interfirm tender offers: information<br />

or synergy. Journal of Financial Economics (11), pp. 182-206.<br />

Braunwarth, K. and Friedl, B. (2010). Towards a financially optimal design of IT services. ICIS 2010<br />

Proceedings, Saint Louis.<br />

Campbell, R., Huisman, R. and Koedijik, K. (2001). Optimal portfolio selection in a Value-at-Risk<br />

framework. Journal of Banking& <strong>Finance</strong> 25.<br />

Christensen, H.A. and Montgomery, C.A. (1981). Corporate economic performance: Diversification<br />

strategy versus market structure. Strategic Management Journal (2:4), pp. 327-343.<br />

Cho, W. and Shaw, M. J. (2009). Does IT Synergy Matter in IT Portfolio Selection? ICIS 2009<br />

Proceedings. Paper 160.<br />

Eckbo, B. E. (1983). Horizontal mergers, collusion, and stockholder wealth. Journal of Financial<br />

Economics (11), pp. 241-274.<br />

Erl,T. (2005). Service-Oriented Architecture: Concepts, Technology & Design, Prentice Hall.<br />

Erradi, A., Anand, S. and Kulkarni, N. (2006). SOAF: An Architectural Framework for Service<br />

Definition and Realization.<br />

Farjoun, M. (1998). The independent and joint effects of the skill and physical bases of relatedness.<br />

Strategic Management Journal (19:7), pp. 611-630.<br />

FSF (2006). Free Software Foundation - TAR. Accessed August 6 th 2011,<br />

http://www.gnu.org/software/tar/tar.html.<br />

Friedl, B. (2011). Zur optimalen Granularität von IT-Services - Eine Analyse relevanter ökonomischer<br />

Einflussfaktoren, 10. Int. Tagung Wirtschaftsinformatik, Zürich.<br />

Github (2011). Github Social Coding. Accessed November 9 th 2011, https://github.com/repositories.


Grigore, M. and Rosenkranz, C. (2011). Increasing the Willingness to Collaborate Online: Analysis of<br />

Sentiment-<strong>Dr</strong>iven Interactions in Peer Content Production. ICIS 2011 Proceedings.<br />

Hill, C.W.L. and Hoskisson, R.E. (1987). Strategy and structure in the multiproduct firm. Academy of<br />

Management Review, (12:2), pp. 331-341.<br />

Huschens, S. and Stahl, G. (2004). Granularität dominiert Korrelation, Risknews 06/04.<br />

Katzmarzik (2011). Product differentiation for SaaS-Providers, BISE, 01/2011, pp. 19-31.<br />

Klose, K., Knackstedt, R. and Beverungen, D. (2007). Identification of Services - A Stakeholder-based<br />

approach to SOA development and its application in the area of production planning, ECIS.<br />

Krafzig, D., Banke, K. and Slama, D. (2004). Enterprise SOA: Service Oriented Architecture Best<br />

Practices, Prentice Hall International.<br />

Lee, J., Seo, Y. and Lee, J. (2007). Method and appartus for merging data objects, European Patent<br />

EP181840A2.<br />

Lichter, H. and Mandl-Striegnitz, P. (1999). Defizite im Software- Projektmanagement - Erfahrungen<br />

aus einer industriellen Studie, Informatique 5.<br />

Lütkebohmert, E. and Gordy, M. (2007). Granularity adjustment for Basel II. No 2007,01, Discussion<br />

Paper Series 2: Banking and Financial Studies, Deutsche Bundesbank, Research Centre.<br />

Markowitz, H. (1952). Portfolio Selection: Journal of <strong>Finance</strong>, p. 77-91.<br />

Millard, D. E., Howard, Y., Abbas, N., Davis, H. C., Gilbert, L., Wills, G. B. and Walters, R. J.<br />

(2009). Pragmatic web service design: An agile approach with the service responsibility and<br />

interaction design method. Computer Science.<br />

Milgrom P. and Roberts, J. (1995). Complementarities and fit Strategy, structure, and organizational<br />

change in manufacturing. Journal of Accounting and Economics (19), pp. 179 -208.<br />

Myers, S.C. (1984). <strong>Finance</strong> theory and financial strategy. Interfaces (4), pp. 26-37.<br />

Pfeifer, T. (2001). Qualitätsmanagement: Strategien, Methoden, Techniken, Hanser Fachb., 3. Aufl.<br />

Rockafellar, R. and Uryasev, S. (1999). Conditional value-at-risk for general loss distributions. Journal<br />

of Banking & <strong>Finance</strong>, Volume 26, Issue 7, Pages 1443-1471<br />

Schirmer, I. and Zimmermann, K. (2008). Visualisierung von Projektportfolios zur Unterstützung des<br />

Architekturmanagements, Universität Hamburg.<br />

Standish (2011). The Standish Group CHAOS Manifesto 2011. Accessed June 9, 2011,<br />

http://standishgroup.com/newsroom/chaos_manifesto_2011.php.<br />

Panzar, J. C. and Willig, R. D. (1981). Economies of scope. American Eco. Rev., (71:2), pp. 268-272.<br />

Tanriverdi, H. and Ruefli, T. W. (2004). The Role of Information Technology in Risk/Return<br />

Relations of Firms. Journal of the Association for IS, Vol. 5, No. 11-12, pp. 421-447.<br />

Tanriverdi, H. (2005). Information Technology Relatedness Knowledge Management Capability, and<br />

Performance of Multibusiness Firms. MIS Quarterly (29:2), pp. 311-334.<br />

Tanriverdi, H. (2006) Performance Effects of Information Technology Synergies in Multibusiness<br />

Firms. MIS Quarterly (30:1), pp. 57-77.<br />

Teece, D.J. (1980). Economies of scope & the scope of the enterprise. Journal of Economic Behavior<br />

and Organization (1:3), pp. 223-247.<br />

Teece, D. J. (1982). Towards an economic theory of the multiproduct firm. Journal of Economic<br />

Behavior and Organization (3), pp. 39-63.<br />

Turek, P., Wierzbicki, A. and Nielek, R. (2010). WikiTeams: Evaluating Teamwork in Wikipedia. In:<br />

Proceedings of the WebSci10.<br />

Weill, P. and Vitale, M. (2002). What IT infrastructure capabilities are needed to implement ebusiness<br />

models? MIS Quarterly, (1:1), pp.17-34.<br />

Wiedemann, A. (2002). Messung von Zinsrisiken mit dem Value-at-Risk-Konzept, WISU 12, pp.<br />

1548-1553.<br />

Wilkens, M. (2001). Basel II- Berücksichtigung von Diversifikationseffekten im Kreditportfolio durch<br />

das Granularity Adjustment, Zeitschrift für das gesamte Kreditwesen, 12; pp. 670-676.<br />

Willig, R. (1978). Technology and market structure. American Economic Review (69), pp. 346-351.<br />

Wolke, T. (2008). Risikomanagement; Oldenbourg Wissenschaftsverlag.<br />

Wierzbicki, A., Turek, P. and Nielek, R. (2010). Learning about team collaboration from Wikipedia<br />

edit history. International Symposium on Wikis and Open Collaboration.


Layer 3:<br />

Customer Management in E-<strong>Finance</strong><br />

(<strong>Prof</strong>. <strong>Dr</strong>. Andreas Hackethal, <strong>Prof</strong>. <strong>Dr</strong>. Bernd Skiera,<br />

<strong>Prof</strong>. <strong>Dr</strong>. Oliver Hinz)<br />

� Bhattacharya U., Andreas H., Simon K., Benjamin L., Steffen M. (2012):<br />

Is unbiased financial advice to retail investors sufficient? Answers from a large<br />

field study<br />

In: Review of Financial Studies, Vol. 25, 975-1032.<br />

� Hackethal A., Michael H., Tullio J. (2012):<br />

Financial advisors: A case of babysitters?<br />

In: Journal of Banking & <strong>Finance</strong>, Vol. 36, 509-524.<br />

� Schmitt P., Skiera B., Van den Bulte C. (2011):<br />

Referral Programs and Customer Value<br />

In: Journal of Marketing, Vol. 75, Jan., 46-59.<br />

� Hinz O., Skiera B., Barrot C., Becker J. (2011):<br />

An Empirical Comparison of Seeding Strategies for Viral Marketing<br />

In: Journal of Marketing, Vol. 75, Nov., 55-71.<br />

� Skiera B., Bermes M., Horn L. (2011):<br />

Customer Equity Sustainability Ratio: A New Metric for Assessing a Firm’s<br />

Future Orientation<br />

In: Journal of Marketing, Vol. 75, May, 118-131.


Is Unbiased Financial Advice to Retail<br />

Investors Sufficient? Answers from a Large<br />

Field Study<br />

Utpal Bhattacharya<br />

Indiana University<br />

Andreas Hackethal<br />

Goethe University, Frankfurt<br />

Simon Kaesler<br />

Goethe University, Frankfurt<br />

Benjamin Loos<br />

Goethe University, Frankfurt<br />

Steffen Meyer<br />

Goethe University, Frankfurt<br />

Working with one of the largest brokerages in Germany, we record what happens when<br />

unbiased investment advice is offered to a random set of approximately 8,000 active retail<br />

customers out of the brokerage’s several hundred thousand retail customers. We find that<br />

investors who most need the financial advice are least likely to obtain it. The investors who<br />

do obtain the advice (about 5%), however, hardly follow the advice and do not improve<br />

their portfolio efficiency by much. Overall, our results imply that the mere availability of<br />

unbiased financial advice is a necessary but not sufficient condition for benefiting retail<br />

investors. (JEL D14, G11, G24, G28)<br />

Hwa is thet mei thet hors wettrien the him self nule drinken? [Who<br />

can give water to the horse that will not drink of its own accord?]<br />

—Oldest English proverb, first recorded in 1175, compiled by<br />

Heywood (1546)<br />

This research would not have been possible without the collaboration of a large German bank. We thank this<br />

bank and all its employees who helped us. For their comments, we also thank John Campbell, Daniel Dorn, Eitan<br />

Goldman, Arvid Hoffmann, Craig Holden, Gur Huberman, Roman Inderst, Narasimhan Jegadeesh, Artashes<br />

Karapetyan, Min Kim, Adair Morse, Antoinette Schoar, Jay Shanken, Andrei Shleifer, Paolo Sodini, Meir Statman,<br />

Noah Stoffman, and Brian Wolfe, as well as the conference and seminar participants at Emory, Frankfurt,<br />

Harvard Business School, Henley, National Chengchi University, National Sun Yat-sen University, National<br />

Taiwan University, Wisconsin, the EFA conference in Stockholm, the European Retail Investment Conference in<br />

Stuttgart, and the FIRS conference in Sydney for their comments. Send correspondence to Utpal Bhattacharya,<br />

Kelley School of Business, Indiana University, Bloomington, IN 47405-1701; telephone: (812) 855-3413; fax:<br />

(812) 855-5875. E-mail: ubhattac@indiana.edu.<br />

c○ The Author 2012. Published by Oxford University Press on behalf of The Society for Financial Studies.<br />

All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.<br />

doi:10.1093/rfs/hhr127 Advance Access publication January 12, 2012<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

There is a large and growing body of literature in household finance documenting<br />

how retail investors make serious investment mistakes by deviating<br />

from the prescriptions of normative finance. 1,2 As Campbell (2006) notes,<br />

a number of potential remedies have sprung up over the years to resolve<br />

these investment mistakes, which include financial education, default options,<br />

and regulation by governments. An important remedy is financial advice<br />

provided by professionals. Providing financial advice is big business. In the<br />

United States, the financial planning and advice industry was estimated at<br />

US$39 billion in 2010. 3 The Investment Company Institute (2007) notes<br />

that over 80% of respondents state that they obtain financial advice from<br />

professional advisors or other sources in the United States. In Germany, a<br />

survey among retail investors indicates that more than 80% of investors consult<br />

a financial advisor. 4<br />

However, financial advice is offered by third parties and may be biased. The<br />

theoretical literature notes that the information asymmetry between investor<br />

and advisor may provide camouflage to advisors who act in their own interests<br />

and to the detriment of their clients. 5 Recent empirical evidence supports the<br />

agency conflict claims formalized in the above theoretical models. 6 A survey<br />

1 This literature is too vast to cover here. We thus offer a sample of important findings. The majority of households<br />

do not even participate in the stock market (Guiso, Haliassos, and Jappelli 2002), despite the large equity<br />

premium that exists (Mehra and Prescott 1985; Dimson, Marsh, and Staunton 2007). The few households<br />

that do participate in equity markets hold underdiversified portfolios (Blume and Friend 1975; Kelly 1995;<br />

Goetzmann and Kumar 2008). Underdiversification, with regard to geographical diversification, is particularly<br />

acute—investors are found to exhibit both a home bias and a preference for local stocks (French and Poterba<br />

1991; Lewis 1999; Cooper and Kaplanis 1994; Huberman 2001; Zhu 2002; Ahearne, Griever, and Warnock<br />

2004; Calvet, Campbell, and Sodini 2007). Research observes inertia that results in individual investors making<br />

insufficient portfolio adjustments in response to general market movements (Agnew, Balduzzi, and Sundén<br />

2003; Calvet, Campbell, and Sodini 2009a; Madrian and Shea 2001). Investors trade too much because they<br />

are overconfident (Odean 1999; Barber and Odean 2000; Deaves, Lüders, and Luo 2003). Investors tend to sell<br />

winners too early and hold on to losers too long, which is an investment mistake called the disposition effect<br />

(Shefrin and Statman 1985; Odean 1998; Frazzini 2006). Investors are fixated on cognitive reference points<br />

(Bhattacharya, Holden, and Jacobsen, forthcoming).<br />

2 Barber and Odean (2000) find that overconfidence leads to substantial return decreases after the deduction of<br />

transaction costs. Looking at the aggregate portfolio of individual Taiwanese investors, Barber et al. (2009)<br />

document an annualized loss of 3.8%. Calvet, Campbell, and Sodini (2007) measure the cost of nonparticipation<br />

and underdiversification and report a substantial loss compared with the world market for Swedish households.<br />

Malkiel (1995) and Carhart (1997) show that even mutual funds underperform the market, net of expenses.<br />

Barras, Scaillet, and Wermers (2010) and Fama and French (2010) find no evidence that the average mutual<br />

fund produces alphas that cover its costs.<br />

3 http://www.ibisworld.com/industry/default.aspx?indid=1316.<br />

4 Two-thirds say that they obtain financial advice from their main bank, whereas about a fifth obtain advice from<br />

an independent financial advisor (DABbank 2004).<br />

5 See, e.g., Bolton, Freixas, and Shapiro (2007), Carlin (2009), Inderst and Ottaviani (2009), Carlin and Gervais<br />

(2009), Stoughton, Wu, and Zechner (2011), and Carlin and Manso (2011).<br />

6 Mutual funds in the United States that are sold through brokered channels underperform. Those funds that<br />

provide higher fees are sold more heavily, which, in turn, negatively affects returns (Bergstresser, Chalmers,<br />

and Tufano 2009; Chen, Kubik, and Hong 2011; Edelen, Evans, and Kadlec 2008). Hackethal, Haliassos, and<br />

Jappelli (2012) find that individual investors whose accounts are run by, or in consultation with, a financial<br />

advisor achieve lower returns. Additionally, in an audit study, Mullainathan, Nöth, and Schoar (2009) conclude<br />

that financial advisors seem to be aggravating the existing biases of their investors. In addition, recommendations<br />

made by research analysts who are compromised through an incentive scheme are shown to have only very<br />

limited potential for value enhancement (Womack 1996; Metrick 1999; Barber and Odean 2001).<br />

976<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

of European Union (EU) members by the CFA Institute (2009) revealed that<br />

64% of respondents believe that the prevailing fee structure is intended to steer<br />

sales instead of serve the customer.<br />

If retail investors make serious investment mistakes, and if investment<br />

advice, which is provided to investors by professional advisors as a means<br />

to ameliorate these mistakes, is conflicted and biased, it would seem that<br />

retail investors would benefit if they could instead receive unbiased and<br />

theoretically sound professional advice. This supply-side cure to improve<br />

the portfolio efficiency of retail investors should work. Regulators certainly<br />

think so. The plethora of financial market reforms that are instituted in major<br />

financial markets—like the Restoring American Financial Stability Act of<br />

2010 (also known as the Dodd–Frank Act) in the United States or the Markets<br />

in Financial Instruments Directive in the EU, which are devised mainly to clean<br />

up the conflicts of interest and improve disclosures in the financial sector—are<br />

essentially supply-side remedies.<br />

This article attempts to answer an important public policy question: will<br />

investors really benefit from only supply-side remedies? To be specific, can<br />

unbiased financial advice steer retail investors toward efficient portfolios?<br />

In this study, we worked with one of the largest brokerages in Germany.<br />

For the first time, the brokerage offered financial advice to a random set<br />

of approximately 8,000 active retail customers out of their several hundred<br />

thousand retail customers. The advice was free and was aimed at improving<br />

portfolio efficiency. The advice was unbiased—i.e., it was free of monetary<br />

incentives for the brokerage and was generated by an algorithm that was<br />

designed to improve portfolio efficiency—and, as we will later document,<br />

the advice does significantly improve portfolio efficiency ex post. We have<br />

demographic data on all retail customers who obtain the advice and who<br />

do not obtain the advice. For both groups, we also possess customers’ daily<br />

trading records for a number of years before the advice is offered and up to<br />

ten months after the advice is offered. We can answer some key questions<br />

because of this, including which types of customers accept the offer, whether<br />

the advice received is followed, whether portfolio efficiency improves for the<br />

average advisee who follows the advice or does not follow the advice, and if<br />

investors who most need the financial advice are least likely to obtain it.<br />

In trying to answer the above questions, this article explores the demand<br />

side of financial advice. To the best of our knowledge, this is the first study to<br />

link the recommendations of advisors with actual customer behavior after the<br />

advice is given.<br />

We arrive at several interesting findings. First, those who accept the offer<br />

(5%) are more likely to be male, older, richer, more financially sophisticated,<br />

and have a longer relationship with the brokerage. Second, of those who accept<br />

the offer, the advice is hardly followed. Third, though portfolio efficiency<br />

hardly improves for the average advisee, it does improve for the average<br />

advisee who follows the advice. Fourth, it seems that investors who most need<br />

977<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

the financial advice are least likely to obtain it. Overall, our results imply that<br />

the mere availability of unbiased and theoretically sound financial advice is a<br />

necessary but not sufficient condition for benefiting retail customers. As the<br />

adage goes, you can lead a horse to water, but you can’t make it drink. 7<br />

The construction of our study allows for a greater understanding of the<br />

factors that contribute to a person opting to obtain and then to follow financial<br />

advice. A probit test gives us a good handle to understand why advice is<br />

not being sought. Such reasons include 1) lack of financial sophistication, as<br />

measured by poor past portfolio performance (has many interpretations); 2) a<br />

desire to not increase tax payments; and 3) lack of familiarity and/or trust, as<br />

measured by the length of relationship with the brokerage. It does seem from<br />

the first result that the clients who most need (least need) the financial advice<br />

are the ones who are least likely to obtain it (most likely to obtain it). This<br />

conclusion is later buttressed by a more formal test.<br />

The results of a regression about why clients do not follow the advice once<br />

they get it is not so extensive. Wealthier investors and those with lower-risk<br />

portfolio values tend to follow the advice more often. Though we have few<br />

positive results, we can rule out many obvious suspects, like trust or financial<br />

sophistication or clients not following advice because the advice asked for a<br />

dramatic increase in investments. We suspect the lack of results, with respect to<br />

the other variables in this test, is due to a lack of power; there is little variation<br />

in the dependent variable: most clients who opt for advice do not follow the<br />

advice. The other reason could be that a systematic cause may not exist to<br />

explain why people do not follow advice. 8<br />

To summarize, the contribution our article makes is to highlight the centrality<br />

of the demand-side problem—unbiased financial advice is useless unless<br />

it is followed—and to recognize the limitations of regulations in dealing with<br />

this demand-side problem. How can regulation convince a person to follow<br />

unbiased and sound financial advice, when we do not completely understand<br />

why people follow advice? Though the answer to this question is beyond the<br />

7 Campbell (2006), and especially Campbell et al. (2011), recognize that consumers need financial protection<br />

because of supply-side problems and also demand-side problems. Using the latest research from behavioral<br />

economics, Campbell et al. (2011) make a very powerful argument that many consumers do not have the ability<br />

to understand complex financial products in an age in which they have to make most of their own financial<br />

decisions. Benartzi and Thaler (2004) document demand-side problems in the context of savings decisions. They<br />

offer possible solutions by taking into account behavioral factors when designing savings plans. They develop a<br />

choice architecture system called “Save More Tomorrow” (SMarT), which is designed to help people commit in<br />

advance to defined contribution increases in pension plans. In an experiment on payday loan borrowers, Bertrand<br />

and Morse (2009) examine the cognitive limitations of these borrowers and focus on solutions. They find that<br />

information disclosure that aims at “de-biasing” is effective.<br />

8 A multitude of reasons are posited in the literature that may explain why people may not follow advice.<br />

These incude bounded rationality (Kahneman 2003) and procrastination that leads to inertia (Samuelson and<br />

Zeckhauser 1988; Laibson 1997; O’Donoghue and Rabin 1999). Other studies, such as Barber and Odean<br />

(2000), find that investors tend to be overconfident, and overconfident individuals tend not to follow advice.<br />

Other influences on the propensity to opt for and follow financial advice may be social interaction (Hong,<br />

Kubik, and Stein 2005) and financial literacy (Christelis, Jappelli, and Padula 2010). More articles are discussed<br />

later.<br />

978<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

scope of this article, our article does cast some doubt on some of the reasons<br />

trotted out to answer this question: noncomprehension of the financial advice,<br />

mistrust of advisor, or simply inertia. 9<br />

The analysis is organized as follows. Section 1 describes our field study,<br />

including details of the offer made to the retail customers by the brokerage<br />

and the particulars of the advice. We explain how the recommendations are<br />

generated and argue that the advice is unbiased and theoretically sound.<br />

Section 2 details the raw data and describes the methodology used to estimate<br />

portfolio efficiency and the degree to which investors follow the advice.<br />

Section 3 gives some revealing, descriptive statistics. Section 4 examines<br />

which retail customers are most likely to choose advice. Section 5 examines<br />

who follows the advice. Section 6 explores the portfolio efficiency of the<br />

customers after receiving the advice. Section 7 analyzes which customers<br />

would most benefit from the advice. Section 8 concludes.<br />

1. Field Study<br />

1.1 Overview<br />

The brokerage we work with was originally founded as a direct bank. Its<br />

focus was on offering brokerage services via telephone and the Internet,<br />

and, over time, it evolved into a full service bank, providing clients with<br />

brokerage and banking services as well as advice on mortgages. However,<br />

the bank never offered investment advice to their clients; all client trades<br />

were self-directed. This changed in 2009. In order to retain existing customers<br />

and attract new ones, the brokerage set out to introduce a financial advisory<br />

business. As a new entrant to the investment advisory market, the brokerage<br />

designed a financial advice model that was distinctly different from those<br />

offered by traditional retail banks. First, the financial advice offered would<br />

not be conflicted; i.e., recommendations would be independent of product<br />

issuers. Second, the financial advice would not be discretionary advice from an<br />

individual advisor but rather recommendations produced by an optimizer that<br />

improves portfolio efficiency. The optimizer would use primarily exchangetraded<br />

funds (ETFs) and mutual funds to increase diversification within and<br />

across asset classes, both domestic and foreign. Third, in order to ensure<br />

and signal the objectivity of its financial advice, the bank would avoid any<br />

9 The same issues are faced in medicine as well. Research on patients’ adherence to medical advice has<br />

been conducted for decades. A meta-analysis of fifty years of research in this field finds that “the average<br />

nonadherence rate is 24.8%.” The reasons for not following a doctor’s advice are that patients think they know<br />

more than the doctor, depression, lack of social support, or simply because they misunderstand or forget what<br />

they have been told. Adherence increases with more circumscribed regimens, as well as education and income,<br />

but not as a function of demographic characteristics (such as gender or age) or the severity of illnesses (DiMatteo<br />

2004). Another finding in the medical literature, which parallels our finding, is that people who most (least) need<br />

to go to their doctor go less (more) often.<br />

979<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

incentive problems by not charging commission on trades that were based on<br />

the recommendations offered. Fourth, during a test phase from which our data<br />

originate, the advice was free of charge.<br />

1.2 Details of the offer<br />

About 8,000 active customers were randomly selected from the brokerage’s<br />

several hundred thousand customers. Active customers are defined as customers<br />

with an account volume of at least e25,000, who have made at least<br />

three trades over the past twelve months and are between eighteen and eighty<br />

years of age. A total of 8,195 customers were offered the advice. In early May<br />

2009, an e-mail was sent to the selected customers’ banking account inboxes<br />

(not private e-mail). In this e-mail, the new advisory service was advertised to<br />

be objective. It was mentioned that 1) the recommendations would be systemgenerated<br />

and independent of product issuers; 2) no commissions, overt or<br />

covert, would be charged for trades that were based on the recommendations;<br />

and 3) the advice would be free during the test phase. Customers were told<br />

that at the end of the test phase, the free advisory service would be terminated<br />

automatically. It was also made clear to the customers that there would be no<br />

obligation to make any transactions that were based on the recommendations<br />

given. Thus, there would be no risk of unintended future commitment for<br />

the customer. If customers did not respond to the offer, a follow-up phone<br />

call was initiated in which an advisor explained the offer again and answered<br />

questions.<br />

1.3 Details of the advice<br />

All customers who opt to receive the free advice from the brokerage are<br />

assigned to our treatment group. All customers who decline the offer form our<br />

control group. Note that this is not a random assignment. However, our basic<br />

empirical methodology—difference in difference—somewhat ameliorates this<br />

shortcoming of our field study.<br />

Every person in the treatment group was contacted by an advisor to<br />

schedule an initial call. This call was used to gather additional demographic<br />

information (e.g., job and household size) and wealth proxies (income and<br />

total financial wealth, including cash and other assets). Risk preferences were<br />

mainly solicited by asking advisees to select between six categories that ranged<br />

from “safe” to “opportunity” as their investment philosophy. There were no<br />

replies for the two most risk-averse categories; we are therefore left with<br />

four levels of risk aversion. Based on demographic data and customer inputs,<br />

the brokerage calculated a risk capacity score that determined the maximum<br />

possible level of risk a client should be exposed to in the recommended<br />

portfolio.<br />

This risk capacity score is the main input for forming customer-specific<br />

recommendations that enhance portfolio efficiency. The customer received<br />

980<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

detailed documentation (see Appendix A for a disguised example of a detailed<br />

recommendation) that included the following information:<br />

• Description of the idea of diversification by investing in different asset<br />

classes and markets<br />

• Explanation of important concepts (e.g., volatility, mean-variance efficiency,<br />

and the Sharpe ratio)<br />

• Analysis of the existing portfolio (historical and expected risk and return<br />

profile)<br />

• Analysis of the recommended portfolio (list of securities and risk and return<br />

profile compared with the existing portfolio)<br />

• Inclusion of client’s requests into recommendations (e.g., securities to<br />

retain)<br />

• Consideration of tax advantages by keeping old investments in the portfolio<br />

• List of trades that are necessary to realize the recommended portfolio<br />

• Fact sheet for each security on the recommended list<br />

In addition to the detailed written documentation that was sent via e-mail, an<br />

advisor also explained the recommendations to the customer over the phone.<br />

1.4 Recommendations<br />

The bank’s recommendations are generated by a mean-variance optimizer that,<br />

based on the original framework of Markowitz (1952), focuses on portfolio<br />

efficiency. The household finance literature indicates that retail investors make<br />

mistakes by holding underdiversified portfolios. This underdiversification is<br />

typically not linked to investment skill (Goetzmann and Kumar 2008). If a<br />

portfolio optimizer improves diversification, it will consequently add value for<br />

a typical retail investor.<br />

It is relatively unimportant which optimization method is applied, as long<br />

as it increases diversification. DeMiguel, Garlappi, and Uppal (2009) show<br />

that more sophisticated techniques are not significantly better than a naïve 1/N<br />

portfolio strategy. To build efficient portfolios, however, it is always desirable<br />

that the expected return estimates are not biased by any past extreme return<br />

realizations. Two precautionary measures are taken. First, the optimizer uses<br />

a shrinkage factor, as proposed by Jorion (1986) and Michaud (1998). The<br />

shrinkage factor is implemented by using a security’s deviation from the longrun<br />

average return of securities with a comparable level of risk. Second, and<br />

more importantly, the optimizer is set up in such a way that it selects from a set<br />

of only eighty securities, predominantly ETFs and/or mutual funds. For such<br />

highly diversified portfolios, Holden (2009) notes that the potential effect of<br />

past idiosyncratic realizations is minimized. Volatilities are estimated by using<br />

981<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

past volatilities. The risk capacity score of the customer is the final input into<br />

the optimizer.<br />

Even though only eighty securities are considered in the basic investment<br />

opportunity set of the optimizer, the potential opportunity set of the optimizer<br />

is much larger because it is capable of considering almost all securities held<br />

by the investors in our sample. Additionally, the optimization is subject to<br />

some constraints and side conditions, such as the maximum weight on an<br />

asset class due to a client’s wishes or risk capacity, maximum number of<br />

securities, minimum weight on a single security, and short-sale constraints.<br />

It is also possible to define other constraints, such as the number of securities<br />

retained from the existing portfolio. As shown in Table A2 in Appendix B,<br />

about 25% of the value of the original portfolio is retained on average. This is<br />

done for two reasons: first, to retain some securities in which customers have<br />

a tax advantage, and second, to increase the chance that the investors will act<br />

on the recommendations. This implies that the recommended portfolios are<br />

different across investors because investors differ in their risk capacity score,<br />

prior portfolio allocations, side constraints, and the point in time at which they<br />

receive their recommendations.<br />

Despite these constraints, Table 1 (panels A–C) shows that the optimizer<br />

is indeed able to improve the diversification of the portfolios along three<br />

important dimensions. First, Table 1, panel A, shows that investments in single<br />

stocks are reduced from 53% to 27%. Clients are advised to invest 67% in<br />

well-diversified ETFs and mutual funds. Recall that the bank has no incentive<br />

to push certain funds, since no commissions are charged. Second, Table 1,<br />

panel B, shows that the average recommended portfolio is more diversified in<br />

different asset classes than is the average existing portfolio before the advice.<br />

The share of equity is reduced from 73% to 59%, while the share of fixed<br />

income and real estate securities is increased from 7% to 23%. Likewise, the<br />

average share of commodities increased from 1% to 13%. Third, Table 1,<br />

panel C, shows that international diversification is strongly enhanced by the<br />

recommendations. Prior to the advice, investors hold about 52% of their equity<br />

in German securities; the recommendations suggest holding 30%.<br />

Table 1, panel D, shows the size of the recommended portfolios compared<br />

with their other portfolios for each advisee. As can be seen, the median<br />

recommended portfolio has exactly the same size as the original portfolio. This<br />

implies that most of our clients are advised not to increase or decrease their<br />

investments in risky assets. This table also shows that the mean recommended<br />

portfolio is 19% larger than the original portfolio. This is because, by default,<br />

the optimizer matched the size of the recommended portfolio with the size of<br />

the advisee’s original portfolio, but the advisee could request a larger (smaller)<br />

size of the recommended portfolio if he or she wanted to invest (divest). Even<br />

those investors who wanted to invest considerably more had sufficient financial<br />

assets to cover this net investment, as shown by the ratio of the recommended<br />

portfolio divided by total financial wealth (median 48%; even the ninety-fifth<br />

982<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 1<br />

Advisees’ original portfolios versus recommended portfolios<br />

Portfolio share<br />

Instrument Original (%) Recommended (%)<br />

Single stocks 53 27<br />

Funds 26 57<br />

ETFs 4 10<br />

Single bonds 4 0<br />

Others 13 6<br />

Total (%) 100 100<br />

Portfolio share<br />

Asset class Original (%) Recommended (%)<br />

Equity 73 59<br />

Fixed income 6 15<br />

Commodities 1 13<br />

Real estate 1 8<br />

Others 19 5<br />

Total (%) 100 100<br />

Equity share<br />

Region Original (%) Recommended (%)<br />

Germany 52 30<br />

Europe 14 24<br />

North America 12 10<br />

Asia Pacific 9 18<br />

World 8 9<br />

Central and eastern Europe 2 7<br />

Other emerging markets 3 2<br />

Total (%) 100 100<br />

Percentile<br />

Standard<br />

Ratio Mean Median 5th 25th 75th 95th deviation N<br />

Recommended portfolio over 1.19 1.00 0.92 0.99 1.04 1.88 0.91 380<br />

original portfolio<br />

Recommended portfolio over 0.90 0.90 0.43 0.75 0.99 1.29 0.43 365<br />

financial wealth with the brokerage<br />

Recommended portfolio over 0.51 0.48 0.15 0.32 0.67 0.97 0.26 318<br />

total financial wealth<br />

Recommended portfolio over 0.28 0.21 0.05 0.12 0.40 0.77 0.23 318<br />

total wealth<br />

Panel A shows the portfolio share by instrument of the advisees’ original portfolio compared with the portfolio<br />

recommended to the advisees at the time of the recommendation. Panel B shows the portfolio share by asset class<br />

of the advisees’ original portfolio compared with the portfolio recommended to the advisees at the time of the<br />

recommendation. Panel C provides a regional breakdown of the equity share for the advisees’ original portfolio<br />

compared with the portfolio recommended to the advisees at the time of the recommendation. Panel D compares<br />

the size of the advisees’ recommended portfolios with the size of other portfolios. The original portfolio of an<br />

advisee is the size of the portfolio held at the time of the recommendation. Financial wealth with the brokerage<br />

is the advisee’s original portfolio plus the value of cash at the brokerage. Total financial wealth is the advisee’s<br />

original portfolio plus the value of cash at the brokerage plus an estimate of financial assets held outside the<br />

brokerage (this estimate is provided by the client). Total wealth is the advisee’s original portfolio plus the value<br />

of cash at the brokerage plus an estimate of financial and nonfinancial assets held outside the brokerage (this<br />

estimate is provided by the client). All sizes are measured in euros at the time of the recommendation. Different<br />

counts of observations are due to data availability of certain variables (five customers had no risky portfolios at<br />

the time of the recommendations).<br />

983<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table 2<br />

Average asset class shares of recommended portfolios by advisees’ risk aversion<br />

Risk aversion<br />

Asset class 1 = Highest 2 3 4 = Lowest<br />

Equity 38 43 53 63<br />

Fixed income 27 20 14 7<br />

Money market 14 15 13 10<br />

Commodities 5 9 10 13<br />

Real estate 14 10 7 5<br />

Other 2 4 3 3<br />

Total 100 100 100 100<br />

N 20 95 172 55<br />

Table 2 shows the average asset class share within advisees’ recommended portfolios for each level of risk<br />

aversion.<br />

percentile of the distribution had enough financial assets to cover the required<br />

net investment with a ratio of 97%) in Table 1, panel D. As our clients may<br />

not follow advice because they are asked to increase their investments in<br />

risky securities, we use the ratio of the recommended and original portfolio<br />

as another independent variable in our test to discover why clients do or do not<br />

choose to follow the advice they opted for. We will find later that this variable<br />

has no significant effect on this decision.<br />

Table 1, panel D, also shows that the median recommended portfolio is 90%<br />

of the financial wealth held with the brokerage (the other 10% is cash), is 48%<br />

of the total financial wealth, and is 21% of total wealth. These large numbers<br />

imply that the risky portfolio investment held in this brokerage is not “play<br />

money.”<br />

Table 2 provides evidence that the optimizer also aligns each client’s risk<br />

capacity with the riskiness of his or her recommended portfolio. From risk<br />

class 1 (highly risk-averse) to risk class 4 (least risk-averse), the share of equity<br />

and commodities monotonically increases, while the share of fixed income,<br />

money market, and real estate investments declines.<br />

Overall, given the constraints, it seems as if the optimizer manages to<br />

recommend well-diversified portfolios to each investor.<br />

1.5 Timeline<br />

The offer was sent to a random sample of 8,195 customers who were drawn<br />

from a population of several hundred thousand customers in early May 2009.<br />

This is our “event date,” which we index as t = 0 on the event timeline.<br />

We collect demographic data on our sample of 8,195 customers. It is an<br />

unbalanced panel. Only 5,952 customers were with the brokerage in September<br />

2005. The other 2,243 joined afterward but before May 2009. The period<br />

between September 2005 and May 2009 is called the “pre-advice” period.<br />

Thus, the pre-advice period is t = −44 months to t = 0. The 8,195 sample<br />

984<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Figure 1<br />

Timeline<br />

The sequence of events in the field study (dates are always at the beginning of the respective month).<br />

customers could join the treatment group between May and October 2009. The<br />

first free recommendation was given on May 13, 2009, and the last customer<br />

opted in during the last week of September 2009. This is why we define the<br />

“post-advice” period as after September 2009. The “post-advice” period lasts<br />

six months, ending on March 31, 2010. Figure 1 illustrates the timeline.<br />

2. Data and Methods<br />

2.1 Data collected<br />

The first part of the data set consists of demographics for the entire random<br />

sample group of 8,195 customers. Table 3 shows the collected data.<br />

These data include gender, age, and microgeographic status, as well as<br />

time-invariant account information, such as the account opening date. The<br />

microgeographic status measures the average wealth level of people living in<br />

Table 3<br />

Data collected<br />

Type of data Type of data Dates available<br />

Client demographics Gender Time-invariant<br />

Date of birth (measure of age) Time-invariant<br />

Microgeographic status (measure of wealth) Time-invariant<br />

Total financial and nonfinancial wealth (advisees only) Time-invariant<br />

Risk aversion Time-invariant<br />

Portfolio characteristics Actual position statements Monthly<br />

Actual transactions Daily<br />

Recommended securities On day of recommendation<br />

Recommended transactions On day of recommendation<br />

Cash On day of recommendation<br />

Number of trades Daily<br />

Account opening date (measure of length of relationship) Time invariant<br />

Market data Carhart (1997) four factors on broad domestic index<br />

−Market factor Weekly<br />

−Small minus big (SMB) Weekly<br />

−High minus low (HML) Weekly<br />

−Momentum factor Weekly<br />

Individual security returns Daily<br />

Table 3 summarizes the data collected during the course of the study. Client demographics and portfolio<br />

characteristics have been provided by the brokerage. The record date is July 2010. Market data are taken from the<br />

Thomson Reuters Datastream. The third column reports the availability of time-series data and their frequency.<br />

985<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

a given micro-area (on a street level). It has nine categories, with nine being<br />

the wealthiest. This variable is provided by a specialized data service provider<br />

that uses several factors (such as house type and size, dominant car brands,<br />

rent per square meter, and unemployment rate) to construct this variable. For<br />

our multivariate tests, we further group this variable into three categories:<br />

low (1–3), medium (4–6), and high wealth (7–9). The account opening date<br />

allows us to compute the length of the relationship between a customer and the<br />

brokerage. The total financial and nonfinancial wealth (e.g., real estate) held<br />

by the client outside the brokerage is an estimate given to us only by clients<br />

who opted to take the advice.<br />

The second part of the data set includes portfolio characteristics for each<br />

customer. Portfolio characteristics include position statements, transactions,<br />

and transfers of holdings from other portfolios. Position statements are on a<br />

security-by-security level and are taken from the beginning of each month.<br />

We obtain the International Security Identification Number (ISIN), the number<br />

of securities held per ISIN, and the respective euro-value for each position.<br />

Transactions and transfers are recorded on a daily basis. For each transaction,<br />

we know the ISIN, trade volume, and transaction price. We use this to calculate<br />

portfolio turnover, as in Barber and Odean (2001), as well as trades per<br />

month. We also have information on the cash accounts of each customer at<br />

t = 0, enabling us to calculate the risky share as the risky portfolio value<br />

divided by financial wealth with the brokerage (risky portfolio value plus cash<br />

value).<br />

Furthermore, for the customers who opted to receive advice, we have data<br />

on when a customer received advice and what exactly the customer was<br />

recommended to buy and sell. This permits us to calculate four different daily<br />

return series for each customer in the treatment group (customers who opted<br />

to obtain advice): actual investment returns in the pre-advice period, actual investment<br />

returns in the post-advice period, buy and hold investment returns in<br />

the post-advice period if the portfolio had not changed from the day before the<br />

advice was given, and investment returns on the recommended portfolio in the<br />

post-advice period if advice had been followed exactly. The data also allow us<br />

to calculate two different daily return series for each customer from the control<br />

group (customers who opted not to obtain advice): actual investment returns<br />

in the pre-advice period and actual investment returns in the post-advice<br />

period.<br />

The third part of the data set contains market data from Thomson Reuters<br />

Datastream. Sample customers hold and trade a total of 46,361 securities<br />

over the observation period. Thomson Reuters Datastream covers 97% of<br />

these securities. A number of our tests use the Carhart (1997) model to<br />

estimate risk-adjusted returns. Therefore, we also produce weekly return<br />

series for the following four factors: the country market factor (MKT),<br />

small minus big (SMB), high minus low (HML), and the momentum factor<br />

(MOM).<br />

986<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

2.2 Return calculations<br />

The brokerage data, in conjunction with the market data from Thomson<br />

Reuters Datastream, allow us to compute daily portfolio returns and daily<br />

holdings on a security-by-security level.<br />

To do this, we first infer daily holdings from monthly position statements,<br />

security transactions, and account transfers. We have end-of-day holdings<br />

for the last day in every month. To obtain the next end-of-day holdings, we<br />

multiply the end-of-day value of each holding by the corresponding price<br />

return (excluding dividends but taking into account any capital actions) for that<br />

security. These holdings are then properly adjusted for any sales, purchases,<br />

and account transfers that occurred on that same day. We repeat this procedure<br />

for every security and investor for each trading day in a given month. The<br />

holdings on the last day of each month are then reconciled with the true<br />

holdings obtained from the brokerage. 10<br />

Second, we compute daily portfolio returns as the weighted average of the<br />

returns of all securities held, purchased, or sold by the investor on that day. We<br />

use total return data (including dividends) for securities without transactions on<br />

that day. For securities that are either purchased or sold, we take into account<br />

exact transaction prices to compute returns. We weight each security’s return to<br />

calculate investors’ daily portfolio returns. All holdings and sales are weighted<br />

by using values in euros on the basis of the previous day’s closing prices. All<br />

purchases are weighted by using the transaction value in euros.<br />

To compute hypothetical returns for portfolios recommended by the broker,<br />

we follow the same procedure and assume that purchase and sale transactions<br />

occur at the price that prevails at the end of the day on which the recommendation<br />

was given. 11<br />

For our regressions, we cumulate all daily portfolio returns into weekly<br />

returns. Portfolio excess returns are weekly portfolio returns minus the riskfree<br />

rate, which we assume to be equal to the three-month EURIBOR. We<br />

regress this excess return on the four factors used by Carhart (1997),<br />

R j,w − R f,w = a j,w + β j,w ∙ � �<br />

Rm,w − R f,w + s j,w ∙ SM Bw<br />

+ h j,w ∙ H M Lw + m j,w ∙ M O Mw + ε j,w, (1)<br />

where R j,w is the return on investor j’s portfolio in week w, R f w is the threemonth<br />

EURIBOR rate in week w, Rm,w is the return in week w on a broad<br />

domestic stock market index (MKT), SMBw and HMLw are the returns for the<br />

size and value-growth portfolios, in accordance with Fama and French (1993),<br />

10 The deviations between inferred and actual holdings are negligible.<br />

11 We can compute net returns for actual transactions because we know all transaction costs. However, since we<br />

cannot compute these net returns for the benchmark “buy and hold” portfolios or the “recommended” portfolios,<br />

all our analysis is done with bid-ask adjusted gross returns (computed at bid and ask prices). Net returns for<br />

actual portfolios are, on average, 1.5% lower per year than their corresponding gross returns.<br />

987<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

in week w, and MOMw is the one-year momentum return from Carhart (1997)<br />

in week w.<br />

The intercept (alpha) in regression Equation (1) is our measure for riskadjusted<br />

portfolio return. For robustness, we also calculate Jensen’s one-factor<br />

alphas for both the domestic stock market as well as the MSCI World in euros.<br />

All results that we report later are robust to using these measures. Note that we<br />

estimate alphas only in the pre-advice stage, where we have forty-four months<br />

of data. We do not estimate alphas in the post-advice period, since we only have<br />

six months of data. We do, however, estimate measures of portfolio efficiency,<br />

as described in the next subsection.<br />

2.3 Measures of portfolio efficiency<br />

We first focus on measuring diversification in a portfolio. To do so, we<br />

use two primary measures—the Herfindahl-Hirschman index (HHI) and the<br />

idiosyncratic variance share—and a secondary measure—home bias. The<br />

HHI is a commonly accepted and simple measure of diversification (Dorn,<br />

Huberman, and Sengmueller 2008; Ivkovic, Sialm, and Weisbenner 2008).<br />

It is calculated by summing the squared portfolio weights of all securities.<br />

Therefore, it follows that the lower the HHI, the better the diversification.<br />

We follow Dorn, Huberman, and Sengmueller (2008) in assuming that if the<br />

security is a fund, the fund consists of one hundred equally weighted positions.<br />

We use idiosyncratic risk in investor portfolios as a measure of diversifiable<br />

risk. For that purpose, we take the variance of the residuals from regression<br />

Equation (1) and, as in Calvet, Campbell, and Sodini (2007), divide it by the<br />

total variance of the dependent variable from the same regression. As this share<br />

increases, the investor bears more diversifiable idiosyncratic risk. Therefore, it<br />

follows that the lower this ratio, the better the diversification. We calculate the<br />

home bias as the percentage of equity in German companies out of total equity.<br />

Therefore, it follows that the lower the home bias, the better the diversification.<br />

We next focus on two portfolio performance metrics: the Sharpe ratio<br />

and a manipulation-proof performance measure (MPPM). The Sharpe ratio<br />

(Sharpe 1966) is a commonly accepted measure of risk-adjusted portfolio<br />

performance. It is calculated by dividing the portfolio excess return by the<br />

portfolio return’s standard deviation. The larger the Sharpe ratio, the better<br />

the portfolio performance. Nevertheless, the Sharpe ratio is subject to several<br />

shortcomings, which Goetzmann et al. (2007) address by proposing a new<br />

Manipulation Proof Performance Measure (MPPM). We follow their MPPM<br />

formula and, similar to other studies (e.g., Deuskar et al. 2011), use the values<br />

2, 3, and 4 for the risk aversion coefficient ρ. We report results for only ρ =<br />

3, as all our results are qualitatively unaltered with ρ = 2 or 4. The larger the<br />

MPPM, the better the portfolio performance. 12<br />

12 The MPPM is not defined for returns of negative 100% (e.g., when an investor holds only one security that<br />

defaults on a specific day and loses all its value). These observations in our data (about 0.001% of the total<br />

number of our observations) are set to missing.<br />

988<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

2.4 Degree of following<br />

To measure the extent to which investors who opted into the advisory model<br />

actually follow the recommendations, we construct a variable that captures the<br />

degree of following the advice:<br />

Degree of following j,d<br />

=<br />

�i=1 N Euro| j, i, d �<br />

actual j, i, drecommended<br />

� i=1<br />

N Euro| j, i, d actual + � i=1<br />

N Euro| j, i, d recommended − � i=1<br />

N Euro| j, i, d actual<br />

� ,<br />

j, i, drecommended<br />

where j denotes the investor, i indicates the specific security, d indexes the<br />

trading day, and Euro is the value in euros that investor j holds in security i on<br />

trading day d. The numerator is the sum in euros of all overlapping securities<br />

(i.e., of those securities that occur in both the actual and the recommended<br />

portfolio). The denominator is the value of the actual portfolio plus the value<br />

of the recommended portfolio, less the overlap. Thus, the degree of following is<br />

actually a ratio between the intersection of the two sets and the union of the two<br />

sets, where the two sets are the actual portfolio and the recommended portfolio.<br />

The ratio can only take on values between zero and one. The ratio is one if a<br />

customer fully follows the advice and zero if the actual and recommended<br />

portfolio do not share a single security.<br />

Table 4 uses a real recommendation to illustrate how this metric works.<br />

The degree of following variable is a measure of how closely advice is<br />

implemented at a particular point in time. We also calculate the change in the<br />

degree of following from the day the advice is given to each day in the period<br />

from t = 5 to t = 11. This change in the degree of following is an exact measure<br />

of the client’s efforts to implement the advice; if the measure is increasing, it<br />

is implied that the advisee is acting in accordance with the recommendations<br />

of the advice.<br />

Our measure considers both the buy and sell sides of following advice.<br />

If a security in the advisee’s portfolio is not included in the recommended<br />

portfolio, this suggests that the investor has been advised to sell this security.<br />

It is possible, however, that the advisee sells a security for reasons other than<br />

following the advice (e.g., liquidity or tax motives). As a check for robustness,<br />

we construct an alternative measure of the degree of following. This measure<br />

considers only the buy side and is simply the share of the recommended<br />

portfolio that the investor holds at any time. Appendix A shows that all results<br />

remain the same qualitatively when using this alternative measure.<br />

3. Descriptive Statistics<br />

3.1 Statistics on clients and portfolios<br />

Table 5 provides summary statistics. It divides the sample group into customers<br />

who opt to obtain the free advice and customers who opt not to obtain the free<br />

(2)<br />

989<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

990<br />

Table 4<br />

Illustration of the degree of following<br />

Original portfolio (t = 0) Recommended portfolio (t = 0) Required Buy Sell Keep<br />

e e action e e e<br />

Deutsche Bank 8,646 −→ Sell 8,646<br />

HSBC Indian Equity Fund 3,622 −→ Sell 3,622<br />

Raiffeisen CEE Equity Fund 2,792 −→ Sell 2,792<br />

HSBC BRIC Equity Fund 1,862 −→ Sell 1,862<br />

Commerzbank 439 −→ Sell 439<br />

E.ON 2,523 E.ON AG 681 −→ Decrease 1,842 681<br />

H&M 3,543 H&M 3,543 −→ Keep 3,543<br />

Comstage ETF EONIA 4,072 −→ Buy 4,072<br />

Schroder ISF Europa Corporate Bond Fund 3,883 −→ Buy 3,883<br />

Allianz Pimco Europazins Bond Fund 3,799 −→ Buy 3,799<br />

Allianz Pimco Corporate Bond Europa Fund 2,434 −→ Buy 2,434<br />

UBS Lux Bond Fund 1,751 −→ Buy 1,751<br />

Grundbesitz Europa Real Estate Fund 1,470 −→ Buy 1,470<br />

Pictet EM Fund 1,247 −→ Buy 1,247<br />

Allianz RCM Small Cap Fund 808 −→ Buy 808<br />

Total 23,426 23,688 19,463 19,202 4,224<br />

Table 4 provides an actual example of the measure that we call the degree of following. Securities E.ON and H&M overlap between the actual and the recommended portfolio. The value<br />

of the overlap is 4,224 euros (e681 in E.ON and e3,543 in H&M). We calculate the degree of following as the overlap between the two portfolios divided by the value of assets in euros in<br />

the actual portfolio plus assets in the recommended portfolio, less the overlap (here: 4,224 / (23,426 + 23, 688 − 4,224) = 10%).<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 5<br />

Summary statistics<br />

Obtain advice Not obtain advice t-test<br />

Data variable Measurement units Mean Median N Mean Median N P-value<br />

Client demographics<br />

Gender Dummy = 1 if male 91.4 100.0 385 81.3 100.0 7,810 .00<br />

Age Years 52.9 52.0 385 49.0 47.0 7,810 .00<br />

Wealth Microgeoraphic status 6.6 7.0 340 6.3 6.0 6,847 .00<br />

Portfolio characteristics<br />

Length of relationship with the bank Years since account opening 9.1 9.7 385 7.4 8.8 7,810 .00<br />

Risky portfolio value at t = 0 e thousands 70.8 46.3 369 45.3 31.6 7,116 .00<br />

Cash at t = 0 e thousands 20.7 10.1 369 21.7 13.2 7,116 .52<br />

Risky share at t = 0 % 75.9 85.4 369 66.1 73.9 7,116 .00<br />

Average trades from t = −44 to t = 0 Trades per months 2.4 1.4 385 1.9 1.0 7,810 .01<br />

Average portfolio turnover from t = −44 to t = 0 %, monthly 5.8 3.6 378 7.5 4.0 7,745 .00<br />

Disposition effect from t = − 44 to t = 0 PGR - PGL 10.2 5.3 385 12.3 2.7 7,810 .13<br />

Share of tax-free assets at t = 0 % 85.8 96.9 369 85.5 98.9 7,081 .82<br />

Portfolio performance measures<br />

Raw returns<br />

From t = −44 to t = 0 (actual portfolios) %, annualized −5.3 −4.7 316 −7.0 −5.4 5,232 .13<br />

From t = 5 to t = 11 (actual portfolios) %, annualized 21.2 0.2 382 17.0 20.2 7,157 .21<br />

From t = 5 to t = 11 (buy and hold portfolios) %, annualized 18.4 18.5 384 .68a From t = 5 to t = 11 (recommended portfolios) %, annualized 24.8 22.6 383 .02a Standard deviation<br />

From t = −44 to t = 0 (actual portfolios) %, annualized 25.8 23.8 316 30.4 26.5 5,232 .00<br />

From t = 5 to t = 11 (actual portfolios) %, annualized 15.0 14.0 382 21.2 17.9 7,157 .00<br />

From t = 5 to t = 11 (buy and hold portfolios) %, annualized 14.5 14.0 384 .00a From t = 5 to t = 11 (recommended portfolios) %, annualized 9.6 8.9 383 .00a Sharpe ratio<br />

From t = −44 to t = 0 (actual portfolios) % −4.1 −4.8 316 −4.0 −4.3 5,231 .69<br />

From t = 5 to t = 11 (actual portfolios) % 21.8 22.5 382 15.8 16.5 7,137 .00<br />

From t = 5 to t = 11 (buy and hold portfolios) % 13.9 18.7 384 .02a From t = 5 to t = 11 (recommended portfolios) % 35.2 35.4 384 .00a (continued)<br />

991<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

992<br />

Table 5<br />

Continued<br />

Obtain advice Not obtain advice t-test<br />

Data variable Measurement units Mean Median N Mean Median N P-value<br />

MPPM<br />

From t = −44 to t = 0 (actual portfolios) % −16.0 −13.3 316 −22.3 −14.6 5,232 .06<br />

From t = 5 to t = 11 (actual portfolios) % 17.3 17.5 382 9.4 16.0 7,157 .03<br />

From t = 5 to t = 11 (buy and hold portfolios) % 14.4 15.0 384 .18a From t = 5 to t = 11 (recommended portfolios) % 22.4 21.2 384 .00a Four-factor alpha<br />

From t = −44 to t = 0 (actual portfolios) %, annualized −6.0 −5.6 316 −8.5 −6.3 5,231 .03<br />

Four-factor beta<br />

From t = −44 to t = 0 (actual portfolios) 0.8 0.8 316 0.9 0.9 5,231 .21<br />

From t = 5 to t = 11 (actual portfolios) 0.5 0.5 381 0.7 0.7 7,080 .00<br />

From t = 5 to t = 11 (buy and hold portfolios) 0.6 0.6 384 .00a From t = 5 to t = 11 (recommended portfolios) 0.4 0.4 384 .00a Portfolio diversification measures<br />

HHI<br />

At t = 0 (original portfolios) % 12.0 7.6 369 20.3 10.8 7,100 .00<br />

At t = 11 (actual portfolios) % 10.4 4.8 377 19.8 9.8 7,014 .00<br />

At t = 11 (buy and hold portfolios) % 9.7 4.7 384 .00a At t = 11 (recommended portfolios) % 2.9 1.6 384 .00a Idiosyncratic variance share<br />

From t = −44 to t = 0 (actual portfolios) % 36.3 30.6 316 39.5 33.9 5,231 .01<br />

From t = 5 to t = 11 (actual portfolios) % 29.6 21.9 381 32.5 23.7 7,080 .03<br />

From t = 5 to t = 11 (buy and hold portfolios) % 30.0 19.6 384 .06a From t = 5 to t = 11 (recommended portfolios) % 21.2 18.1 384 .00a (continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 5<br />

Continued<br />

Obtain advice Not obtain advice t-test<br />

Data variable Measurement units Mean Median N Mean Median N P-value<br />

Home bias<br />

At t = 0 (original portfolios) % Of equity in Germany 51.2 51.7 369 55.2 58.2 7,116 .03<br />

At t = 11 (actual portfolios) % Of equity in Germany 43.9 41.3 378 51.1 52.2 7,177 .00<br />

At t = 11 (buy and hold portfolios) % Of equity in Germany 45.6 42.5 384 .00a At t = 11 (recommended portfolios) % Of equity in Germany 30.3 28.7 384 .00a aActual portfolios used for non-advised clients<br />

Table 5 reports summary statistics on client demographics, portfolio characteristics, portfolio performance measures, and portfolio diversification measures. The columns “Obtain advice”<br />

and “Not obtain advice” present means, medians, and number of observations in regard to the respective clients in each group. The last column reports p-values of a t-test on a difference<br />

of means. Client demographics comprise statistics on the share of male clients (Gender), the age of clients (Age), and the wealth of a client measured by the microgeographic status rating,<br />

one through nine, by an external agency (Wealth). Portfolio characteristics comprise statistics on the number of years the client has been with the bank (Length of relationship), the risky<br />

portfolio value of the customer (Risky portfolio value at t = 0), the amount of cash held with this brokerage (Cash at t = 0), the proportion of risky assets held with this brokerage (Risky<br />

share at t = 0), the number of trades per month (Average trades from t = −44 to t = 0), the average portfolio turnover per month (Average portfolio turnover from t = − 44 to t = 0), the<br />

difference between the proportion of realized gains and losses (Disposition effect from t = − 44 to t = 0), and the proportion of tax-free assets (Share of tax-free assets at t = 0). Portfolio<br />

performance measures comprise statistics on raw returns, standard deviations, Sharpe ratios, Manipulation Proof Performance Measures (MPPM), four-factor alphas, and four-factor betas<br />

for the period before the advice was offered, t = −44 to t = 0, and the period after the advice was offered, t = 5 to t = 11. Portfolio diversification measures comprise statistics on the<br />

Herfindahl-Hirschmann index (HHI), the idiosyncratic variance share (Idiosyncratic variance share), and the share of domestic equity (Home bias). Alpha, beta, and idiosyncratic risk share<br />

stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany.<br />

993<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

advice. P-values of t-tests from our tests for the equality of variables across<br />

these two groups are provided in the last column.<br />

Table 5 shows that 91% of the customers who accepted the offer are male,<br />

compared with 81% in the control group. The mean age is also slightly<br />

higher (52.9 years vs. 49.0 years), as is the wealth level measured by the<br />

microgeographic status (6.6 vs. 6.3). This indicates that the customers who<br />

accept the offer are more likely to be male, older in age, and richer. The<br />

customers who accept the offer have had a longer relationship with the bank<br />

(9.1 years vs. 7.4 years). Portfolio characteristics are also significantly different<br />

for the two subgroups of our sample. The customers who accept the offer have<br />

a higher-risk portfolio value at t = 0 (e70,800 vs. e45,300), a higher share<br />

of risky assets (75.9% vs. 66.1%), more trades per month (2.4 vs. 1.9), and<br />

lower portfolio turnover (5.8% vs. 7.5%). 13 We also calculate the disposition<br />

effect for each investor during the pre-advice period by applying the Odean<br />

(1998) method. Differences between advisees and non-advisees exist but are<br />

not significant.<br />

For those customers who have an account from September 2005 to May<br />

2009—the pre-advice period—we calculate average daily returns of their<br />

investment portfolios, the standard deviations of these returns, and four-factor<br />

alphas. The raw returns are not significantly higher for the customers who<br />

accept the offer than they are for the customers who do not accept the offer.<br />

The standard deviations of returns are significantly lower for the customers<br />

who accept the offer than they are for the customers who do not accept the<br />

offer. The alphas are significantly higher for the customers who accept the<br />

offer than they are for customers who do not accept the offer, though both their<br />

alphas are significantly negative. The Sharpe ratios and the betas are similar.<br />

However, MPPM is significantly higher for the customers who accept the offer<br />

than it is for customers who do not accept the offer, though for both groups<br />

the MPPMs are negative. All of this evidence suggests that the customers who<br />

accept the offer are likely to be more financially sophisticated.<br />

We also notice that diversification (as measured by lower HHI, lower<br />

idiosyncratic risk share, and lower home bias) is significantly higher for the<br />

customers who accept the offer than it is for the customers who do not accept<br />

the offer. This further confirms that the customers who accept the offer are<br />

likely to be more financially sophisticated. All the above results are confirmed<br />

in Section 4 by multivariate tests.<br />

It is important to mention that the largely negative alpha estimates in the<br />

pre-advice period tell us that all our customers, regardless of whether they<br />

accept the offer, significantly underperform the benchmark index. Similarly,<br />

13 In comparison with official statistics provided by Deutsche Bundesbank (2010) and Deutsches Aktieninstitut<br />

(2009), the investors in our sample have about the same age (fifty years) but are more likely to be male and<br />

richer. The larger portfolio and large cash values signal that these brokerage accounts do not represent “play<br />

money” (Goetzmann and Kumar 2008). In line with Calvet, Campbell, and Sodini (2007), our investors also<br />

hold portfolios with high idiosyncratic volatility shares.<br />

994<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

high idiosyncratic risk shares, high HHIs, and high home bias show significant<br />

potential for improvement of diversification. Regardless of whether investors<br />

are self-directed or follow some outside advice, they could benefit from<br />

unbiased and theoretically sound advice, especially the ones who were doing<br />

relatively worse.<br />

We notice that the improvement in actual raw returns from the pre-advice<br />

period to the post-advice period for the customers who accept the offer (−5.3%<br />

to 21.2%) does not seem to be much different from the customers who do<br />

not accept the offer (−7.0% to 17.0%). The drop in standard deviations from<br />

the pre-advice period to the post-advice period for the customers who accept<br />

the offer (25.8% to 15.0%) also does not seem to be much different from the<br />

customers who do not accept the offer (30.4% to 21.2%). A multivariate test<br />

later confirms that obtaining advice does not improve diversification: there is<br />

no significant decrease in HHI or the idiosyncratic risk of advisees’ portfolios<br />

compared with the non-advisees’ portfolios. The results, with respect to<br />

portfolio performance, are mixed: there is a slight increase in Sharpe ratio but<br />

no increase in MPPM of advisees’ portfolios compared with the non-advisees’<br />

portfolios. Taken together, this suggests that the average advisee does not much<br />

benefit from the advice.<br />

There are several explanations to answer the questions of why the average<br />

advisee does not much benefit from the advice. It could be that the advice is not<br />

sound or that the average advisee does not follow the advice. To check whether<br />

the advice is sound, we perform two simple univariate tests: we compare the<br />

recommended portfolios with the investors’ actual portfolios and their buy and<br />

hold portfolios.<br />

Table 5 shows these results. The recommended portfolios perform much better<br />

than the actual portfolios of the advisees in the post-advice period: a return<br />

of 24.8% vs. 21.2%, a standard deviation of 9.6% vs. 15.0%, a Sharpe ratio of<br />

35.2% vs. 21.8%, an MPPM of 22.4% vs. 17.3%, an HHI of 2.9% vs. 10.4%,<br />

an idiosyncratic risk share of 21.2% vs. 29.6%, and a home bias of 30.2% vs.<br />

43.9%. It is also noteworthy that the recommended portfolios perform much<br />

better than the buy and hold portfolios of the advisees in the post-advice period:<br />

a return of 24.8% vs. 18.4%, a standard deviation of 9.6% vs. 14.5%, a Sharpe<br />

ratio of 35.2% vs. 13.9%, an MPPM of 22.4% vs. 14.4%, an HHI of 2.9%<br />

vs. 9.7%, an idiosyncratic risk share of 21.2% vs. 30.0%, and a home bias of<br />

30.2% vs. 45.5%. These results suggest that the financial advice was sound and<br />

that the average advisee does not follow the advice. If he or she had followed<br />

the recommendations, he or she would have improved his or her investment<br />

performance. Formal tests later conducted confirm all these initial findings.<br />

3.2 Statistics on the degree of following the advice<br />

We now give descriptive statistics on the measure of degree of following to give<br />

a sense of how many investors follow the advice once they elect to receive it.<br />

995<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Figure 2<br />

Initial degree of following<br />

Panel A: The cross-sectional distribution of the degree of following when the first recommendation was given.<br />

Panel B: The cross-sectional distribution of the average degree of following between time t = advice start and<br />

t = 11 months. Panel C: Increase in degree of following from the time of the first recommendation to t = 11.<br />

Figure 2, panel A, gives the distribution of the degree of following on the<br />

day the recommendation is received by an investor. This figure shows that the<br />

recommended portfolio is very different from the average advisee’s existing<br />

portfolio. For about one in five investors, there is essentially no overlap, and<br />

for about half of them, overlap is less than 20%. In fact, no one’s existing<br />

portfolio coincides with his or her recommended portfolio.<br />

Figure 2, panel B, gives the distribution of the average degree of following<br />

between t = advice start through t = 11 months, which is the “post-advice”<br />

period, for investors who choose to receive advice. Compared with Figure 2,<br />

panel A, this figure indicates that some mass of the distribution shifted to the<br />

right. This implies that some investors follow the advice in the “post-advice”<br />

period and increase the degree of following. However, the distributions shown<br />

in Figure 2, panels A and B, are not very different from each other, suggesting<br />

that, in fact, few investors follow the advice.<br />

Figure 2, panel C, gives the distribution of the increase in the degree of<br />

following from the date the investor receives the advice to t = 11. As the big<br />

mass is at zero, it tells us that most investors do not follow the advice. In<br />

fact, some mass is in the negative zone, indicating that some investors even<br />

996<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

disregard the advice, i.e., sell securities they are recommended to keep and/or<br />

buy securities that are not recommended. However, more investors follow the<br />

advice and then disregard it. (Though, there are relatively few investors who do<br />

either of these.) Figure 2, panel C, looks similar when we measure the increase<br />

in the degree of following from the date of the recommendation to ten days<br />

later, twenty days later, thirty days later, or use the average increase over the<br />

entire post-advice period.<br />

Table 6 gives the descriptive statistics on distributions of the degree of<br />

following advice taken at different points in time.<br />

We see in this table that the mean degree of following is 15.6% on the day of<br />

the recommendation for the average advisee. This means that, on average, the<br />

majority of the positions in the original portfolio have to be changed. This is<br />

not an onerous task because the average advisee’s portfolio already experiences<br />

an annual turnover of approximately 70%.<br />

For the average advisee, the mean degree of following increases to 21.6%<br />

ten days after the advice, to 24.4% twenty days after the advice, and to 25.4%<br />

thirty days after the advice. However, the mean degree of following is lower<br />

(16.4%) at the end of the post-advice period, suggesting that, although they<br />

may initially act in accordance with the advice, they do not stick to it. This also<br />

explains why the mean degree of following for the entire post-advice period is<br />

only 21.1%. An important statistic in this table is the number of investors who,<br />

at least partially, follow the advice. Of the 385 who opt to receive advice, 260<br />

(385 – 125) do not follow it, as measured at the end of the post-advice period.<br />

The 125 investors who do follow the advice, as can be seen in Figure 2, panel<br />

C, follow it very little.<br />

Our conclusions from Figure 2 and Table 6 are straightforward. Recommended<br />

portfolios are very different from the average advisee’s actual<br />

portfolios, and the average advisee does not much heed the financial advice.<br />

We do not believe that this is due to transaction costs because, as previously<br />

stated, investors frequently turn over their portfolios.<br />

Table 6<br />

Summary statistics for the degree of following<br />

t =11 months Average over<br />

(end of measurement<br />

measurement period (t = 5 to<br />

t = Advice start t = 10 days t = 20 days t = 30 days period) t = 11 months)<br />

Degree of following<br />

Mean (%) 15.6 21.6 24.4 25.4 16.4 21.1<br />

Median (%) 10.8 14.2 15.5 16.2 10.4 14.7<br />

Number of followers n/a 98 141 158 125 157<br />

Observations 385 385 385 385 385 385<br />

Table 6 reports summary statistics for the degree of following. The table reports the mean degree of following,<br />

the median degree of following, the number of followers, and observations for six distinct time periods (upon<br />

receiving the first recommendation, after ten days, after twenty days, after thirty days, and at the end of the<br />

measurement period and the average over the entire measurement period). These time intervals are specific to<br />

each investor in the sample. Number of followers is defined as the number of investors who increase their degree<br />

of following until that point in time, at least marginally.<br />

997<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

4. Who Chooses to Obtain Advice?<br />

By October 2009, a total of 385 customers, out of the 8,195 customers that<br />

were offered free and unbiased financial advice, elected to accept the offer.<br />

This constitutes a little less than 5% of the customers who were contacted.<br />

A total of thirty-eight investors joined in May 2009, 146 in June 2009, seventythree<br />

in July 2009, and the remaining 128 joined later.<br />

We now formally examine those who choose to receive advice. Table 7<br />

reports the results of a probit test, where the dependent variable is set to one<br />

if a client opted to receive financial advice and zero otherwise. 14 We make the<br />

following conclusions.<br />

Though our sample is predominantly male, older, and rich, the clients who<br />

opt to receive free financial advice are more likely to be male, older, richer<br />

(to be precise, the effect comes from low-wealth clients who choose not to<br />

obtain advice), and have a higher risky share in their portfolio. They are more<br />

active traders, as measured by the number of trades per month, but have less<br />

portfolio turnover, relative to their risky portfolio values. Their alphas are<br />

higher, their HHI are lower, their home bias is lower, and their disposition<br />

effect bias is lower. Their long-term raw returns are higher, but their short-term<br />

raw returns are lower. They also have more experience with the brokerage,<br />

where experience is measured by the length of the relationship between the<br />

client and the brokerage. A lower share of tax-free assets, especially when<br />

controlling for diversification, is positively correlated with opting for advice. 15<br />

Age and wealth are linked to financial sophistication in the literature (Calvet,<br />

Campbell, and Sodini 2007; Calvet, Campbell, and Sodini 2009b), though<br />

being male is not linked to financial sophistication (Barber and Odean 2001).<br />

Higher alphas, lower HHI, lower home bias, and lower disposition effect bias<br />

are linked to financial sophistication. So, why do the financially sophisticated<br />

opt for advice? It is possible that these investors became financially sophisticated<br />

because of financial literacy training—see Carlin and Robinson (2010)—<br />

and so they have a positive view of advice. A second hypothesis is plausible:<br />

elderly and financially more sophisticated men are more likely to be defrauded<br />

(NASD 2006) or older investors are more likely to make erroneous financial<br />

investment decisions (Agarwal et al. 2009); opting for financial advice by these<br />

groups could also be interpreted as an attempt to address these disadvantages.<br />

14 In order to address the issue of commonality among the recommendations, all our results in Tables 7 and 8 are<br />

also estimated by using a cluster robust regression analysis, with risk class being the cluster variable. All nonadvised<br />

clients were grouped into one class. Results remain qualitatively unaltered. We also notice that investors<br />

opted for the advice at different points in time. To deal with this, we also use clustered standard errors in Tables<br />

7 and 8, with week of opting being the cluster variable. Again, results remain qualitatively unaltered.<br />

15 Poterba and Samwick (2003) suggest that taxes may influence an investor’s trading decisions. A tax law change<br />

in Germany made it favorable to hold assets that were in existing portfolios at the end of 2008 because they<br />

would remain tax-free if held for more than one year. We calculate the share of tax-free assets in the portfolio<br />

at the time of the offer. As this share increases, propensity to seek financial advice may decrease. However,<br />

advisees sell, on average, 57.5% of their tax-free assets until the end of our measurement period. So, they lose<br />

most of their tax advantage.<br />

998<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 7<br />

Who opts for advice? A probit test<br />

Dummy advice<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Dummy male 0.361∗∗∗ 0.278∗∗∗ 0.285∗∗∗ 0.288∗∗∗ 0.291∗∗∗ 0.290∗∗∗ 0.257∗∗∗ (0.000) (0.001) (0.002) (0.002) (0.000) (0.000) (0.007)<br />

Age 0.009∗∗∗ 0.008∗∗∗ 0.011∗∗∗ 0.010∗∗∗ 0.008∗∗∗ 0.008∗∗∗ 0.011∗∗∗ (0.000) (0.001) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Dummy low wealth −0.255∗ −0.267∗∗ −0.177 −0.183 −0.264∗ −0.271∗∗ −0.200<br />

(0.059) (0.047) (0.213) (0.197) (0.050) (0.043) (0.181)<br />

Dummy high wealth 0.048 0.062 0.047 0.047 0.065 0.067 0.030<br />

(0.353) (0.233) (0.413) (0.417) (0.217) (0.201) (0.615)<br />

Log portfolio value (t =0) 0.144∗∗∗ 0.130∗∗∗ 0.130∗∗∗ 0.131∗∗∗ 0.102∗∗∗ 0.130∗∗∗ 0.138∗∗∗ (0.000) (0.001) (0.004) (0.005) (0.008) (0.001) (0.002)<br />

Length of relationship 0.063∗∗∗ 0.098∗∗∗ 0.098∗∗∗ 0.068∗∗∗ 0.064∗∗∗ 0.102∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Trades per month (t = −44 to t =0) 0.042∗∗∗ 0.043∗∗∗ 0.042∗∗∗ 0.028∗∗∗ 0.037∗∗∗ 0.042∗∗∗ (0.000) (0.000) (0.000) (0.008) (0.000) (0.001)<br />

Portfolio turnover (t = −44 to t =0) −1.104∗∗ −1.042∗ −1.030∗ −0.450 −0.966∗∗ −0.589<br />

(0.010) (0.080) (0.095) (0.310) (0.024) (0.427)<br />

Disposition effect (t = −44 to t =0) −0.163∗ −0.208∗∗ −0.203∗∗ −0.204∗∗ −0.154∗ −0.208∗∗ (0.065) (0.031) (0.035) (0.027) (0.079) (0.039)<br />

Risky share (t =0) −0.103 −0.192 −0.201 −0.082 −0.103 −0.268∗∗ (0.353) (0.115) (0.101) (0.468) (0.356) (0.029)<br />

Share of tax-free assets (t = 0) −0.194∗ −0.219∗ −0.235∗ −0.287∗∗ −0.217∗ −0.223<br />

(0.085) (0.094) (0.075) (0.017) (0.055) (0.151)<br />

Alpha (t = −44 to t = 0) 21.836∗∗ (0.012)<br />

Idiosyncratic variance share (t = −44 to t = 0) −0.189<br />

(0.241)<br />

HHI (t = 0) −0.883∗∗∗ (0.000)<br />

(continued)<br />

999<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1000<br />

Table 7<br />

Continued<br />

Dummy advice<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Home bias (t =0) −0.197∗∗∗ (0.008)<br />

Long-term raw return (t = −44 to t = −1) 27.917∗∗ (0.036)<br />

Short-term raw return (t = −1 to t =0) −2.750∗ (0.100)<br />

Constant −3.937∗∗∗ −3.923∗∗∗ −4.327∗∗∗ −4.266∗∗∗ −3.513∗∗∗ −3.850∗∗∗ −4.287∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Observations 7,450 7,449 5,520 5,520 7,432 7,449 5,056<br />

Pseudo-R2 0.0381 0.0668 0.0677 0.0664 0.0778 0.0691 0.0682<br />

Table 7 reports probit estimates of the participation in the advisory model offered by the brokerage. Clients are set equal to one once they obtain the advice. For the estimation of the<br />

probit model, we include the following independent variables: a dummy that is equal to one if a client is male (Dummy male), the age of a client (Age), a dummy that is equal to one if a<br />

client falls into categories one to three of a microgeographic status rating by an external agency (Dummy low wealth), a dummy that is equal to one if a client falls into categories seven to<br />

nine of the microgeographic status (Dummy high wealth), the risky portfolio value of the customer (Log portfolio value), the number of years the client has been with the bank (Length of<br />

relationship), the number of trades per month (Trades per month), the average portfolio turnover per month (Portfolio turnover), the difference between the proportion of realized gains and<br />

losses (Disposition effect), the proportion of risky assets in the account (Risky share), the proportion of tax-free assets (Share of tax-free assets), the weekly alpha of a particular customer<br />

before opting for financial advice (Alpha), the idiosyncratic variance share (Idiosyncratic variance share), the Herfindahl-Hirschman index (HHI), the share of domestic equity (Home<br />

bias), the raw return from the beginning of observation until one month prior to the offer of financial advice (Long-term raw return), and the raw return of the month before the offer of<br />

financial advice (Short-term raw return). Alpha and idiosyncratic risk share stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in<br />

parentheses. Pseudo-R2 values and number of observations are reported. *** denotes significance at the 1% level or less, ** significance at 5% or less, and * significance at 10% or less.<br />

Heteroscedasticity robust standard errors are used. Standard errors shown are not clustered, but results remain qualitatively unaltered when clustering them by advice week or risk aversion.<br />

Different counts of observations are due to data availability of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

It is also possible that it is not financial sophistication but a third hypothesis<br />

that explains the results: regret. Investors who have performed poorly in the<br />

past have regret. They are reluctant to obtain advice and to acknowledge past<br />

mistakes. They prefer inaction (ignore advice and be wrong, which comes<br />

at a lower psychic cost) to action (obtain advice, change portfolio, and risk<br />

being wrong, which comes at a higher psychic cost). Carlin and Robinson<br />

(1999) have a good example of this happening in Las Vegas. This regret<br />

hypothesis is, however, weakened by our finding that investors with lower<br />

short-term returns opt for advice more often. Furthermore, we computed a<br />

variable, proportion of paper losses, in an advisee’s portfolio when the advice<br />

was offered, and this variable seemed to have no effect. A fourth hypothesis<br />

could be the “once bitten, twice shy” hypothesis: investors who had followed<br />

past advice and had done well want to receive more advice, whereas investors<br />

who had followed past advice and had not done well do not want to receive<br />

more advice. However, any past advice received by these investors could not<br />

have come from this brokerage, as this is the first time that the brokerage<br />

has offered advice. A fifth hypothesis could be that our clients are optimistic,<br />

perhaps extremely optimistic. Puri and Robinson (2007) show that optimists<br />

invest more in individual stocks, and extreme optimists display imprudent<br />

financial behavior (like not wanting advice). While this could be true, we<br />

do not have any psychological metrics on our clients and cannot test this<br />

hypothesis. A sixth hypothesis could be that investors receive advice from<br />

other sources. Even if they receive outside advice, the outside advice must<br />

either be bad or not followed because we find portfolios to be largely inefficient<br />

before, as well as after, the advice is offered. A seventh hypothesis could be<br />

that the offer of free and unbiased financial advice is just regarded as spam.<br />

Again, the evidence suggests otherwise. The e-mail that contained the offer<br />

was an official message sent to the inbox of the banking account. This inbox<br />

is never crowded because these e-mails have a predetermined maturity. Also,<br />

if there is no reaction to the initial e-mail, a personal phone call follows.<br />

To summarize, we do have a fairly good handle on why advice is not being<br />

sought: lack of financial sophistication, as measured by poor past portfolio<br />

performance (has many interpretations), a desire to not increase tax payments,<br />

and lack of experience, familiarity, and/or trust, as measured by the length<br />

of relationship with the brokerage. It does seem from the first result that the<br />

clients who most need (do not need) the financial advice are the least likely to<br />

obtain it (most likely to obtain it). We will explore this hypothesis more fully<br />

in Section 7.<br />

5. Who Chooses to Follow the Advice If They Obtain It?<br />

We now formally examine who chooses to follow the advice once they obtain<br />

it. Recall that advisees rarely follow the advice, and when they do, it is only<br />

followed to a small extent. Table 8 reports the results of an ordinary least<br />

1001<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table 8<br />

Who chooses to follow advice?<br />

1002<br />

Increase in degree of following<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Dummy male −0.080 −0.076 −0.055 −0.062 −0.079 −0.080 −0.048<br />

(0.122) (0.131) (0.298) (0.238) (0.119) (0.115) (0.374)<br />

Age 0.000 0.000 0.000 0.000 0.000 0.000 0.000<br />

(0.535) (0.826) (0.995) (0.967) (0.872) (0.925) (0.797)<br />

Dummy low wealth −0.041∗∗ −0.054∗∗ −0.063∗∗∗ −0.059∗∗ −0.053∗∗ −0.052∗∗ −0.062∗∗ (0.046) (0.031) (0.007) (0.018) (0.035) (0.049) (0.011)<br />

Dummy high wealth 0.045∗∗ 0.044∗ 0.038 0.038 0.044∗ 0.043∗ 0.033<br />

(0.039) (0.051) (0.120) (0.115) (0.051) (0.055) (0.182)<br />

Log portfolio value (t = 0) −0.060∗∗∗ −0.065∗∗∗ −0.060∗∗∗ −0.056∗∗∗ −0.064∗∗∗ −0.066∗∗∗ −0.056∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Length of relationship 0.003 0.006 0.006 0.002 0.003 0.005<br />

(0.398) (0.156) (0.188) (0.459) (0.425) (0.257)<br />

Trades per month (t = −44 to t = 0) 0.001 0.002 0.003 0.002 0.002 0.003<br />

(0.751) (0.636) (0.408) (0.614) (0.618) (0.528)<br />

Portfolio turnover (t = −44 to t = 0) −0.025 −0.052 −0.186 −0.054 −0.051 −0.304<br />

(0.906) (0.842) (0.470) (0.806) (0.811) (0.413)<br />

Disposition effect (t = −44 to t = 0) −0.027 −0.013 0.002 −0.026 −0.030 0.032<br />

(0.579) (0.798) (0.960) (0.591) (0.538) (0.518)<br />

Risky share (t = 0) 0.002 −0.014 −0.018 −0.001 0.003 −0.002<br />

(0.973) (0.753) (0.684) (0.976) (0.945) (0.970)<br />

Share of tax-free assets (t = 0) 0.051 0.054 0.084 0.057 0.057 0.083∗<br />

(0.240) (0.272) (0.126) (0.200) (0.188) (0.056)<br />

Recommended portfolio over original portfolio −0.003 −0.002 −0.003 −0.005 −0.003 0.014<br />

(0.829) (0.841) (0.830) (0.728) (0.785) (0.464)<br />

Alpha (t = −44 to t = 0) 6.160<br />

(0.155)<br />

Idiosyncratic variance share (t = −44 to t = 0) 0.089<br />

(0.152)<br />

HHI (t = 0) 0.061<br />

(0.534)<br />

(continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 8<br />

Continued<br />

Increase in degree of following<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Home bias (t = 0) 0.031<br />

(0.382)<br />

Long-term raw return (t = −44 to t = −1) 8.195<br />

(0.169)<br />

Short-term raw return (t = −1 to t = 0) −0.872<br />

(0.401)<br />

Constant 0.754∗∗∗ 0.752∗∗∗ 0.670∗∗∗ 0.580∗∗∗ 0.741∗∗∗ 0.751∗∗∗ 0.594∗∗∗ (0.000) (0.000) (0.000) (0.001) (0.000) (0.000) (0.001)<br />

Observations 369 365 313 313 365 365 297<br />

R2 0.132 0.139 0.138 0.139 0.141 0.141 0.131<br />

Table 8 reports OLS estimates of the coefficients related to an increase in the degree of following measure from the date the investor received advice to t = 20 days (Increase in degree<br />

of following). For the estimation of the model, we include the following independent variables: a dummy that is equal to one if a client is male (Dummy male), the age of a client (Age),<br />

a dummy that is equal to one if a client falls into categories one to three of a microgeographic status rating by an external agency (Dummy low wealth), a dummy that is equal to one if a<br />

client falls into categories seven to nine of the microgeographic status (Dummy high wealth), the risky portfolio value of the customer (Log portfolio value), the number of years the client<br />

has been with the bank (Length of relationship), the number of trades per month (Trades per month), the average portfolio turnover per month (Portfolio turnover), the difference between<br />

the proportion of realized gains and losses (Disposition effect), the proportion of risky assets in the account (Risky share), the proportion of tax-free assets (Share of tax-free assets), a ratio<br />

of the required net investment (Recommended portfolio/original portfolio), the weekly alpha of a particular customer before opting for financial advice (Alpha), the idiosyncratic variance<br />

share (Idiosyncratic variance share), the Herfindahl-Hirschman index (HHI), the share of domestic equity (Home bias), the raw return from the beginning of observation until one month<br />

prior to the offer of financial advice (Long-term raw return), and the raw return of the month before the offer of financial advice (Short-term raw return). Alpha and idiosyncratic risk share<br />

stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number of observations are reported. *** denotes<br />

significance at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used. Standard errors shown are not clustered, but<br />

results remain qualitatively unaltered when clustering them by advice week or risk aversion. Different counts of observations are due to data availability of certain variables (see Table 5);<br />

results are robust to using the lowest common denominator.<br />

1003<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

squares (OLS) regression, where the dependent variable is the increase in the<br />

degree of following the advice. We find that wealthier investors (as measured<br />

by the microgeographic status) and investors with lower risky portfolio values<br />

tend to follow the advice more. Interestingly, though investors with higher<br />

portfolio value opt for the advice, investors with lower portfolio value follow<br />

it more. One interpretation of this seeming contradiction is that people with<br />

lower portfolio value have, among the richer clients, relatively less experience<br />

with risky assets and may therefore follow advice more.<br />

The more interesting observation in Table 8 is the set of nonresults. None<br />

of the other variables have a significant effect on the degree to which advice is<br />

followed, including the disposition effect. The next paragraph discusses more<br />

nonresults.<br />

First, it is conceivable that investors do not pay attention to advice because<br />

they consider their accounts in the brokerage as just “play money.” The<br />

evidence suggests otherwise. As can be seen in Table 1, panel D, the median<br />

original portfolio of the client has the same size as his median recommended<br />

portfolio which, in turn, is about 48% of his total financial wealth. This<br />

suggests that we are looking at investors’ main accounts and that their financial<br />

wealth is predominantly held by this brokerage. 16 Additionally, in comparison<br />

with official statistics provided by Deutsche Bundesbank (2010) and Deutsches<br />

Aktieninstitut (2009), the average investor in our sample has a much higher<br />

account value than the average retail investor in the population.<br />

Second, it could be that the advice is not followed because investors are<br />

asked to increase their investments in risky assets. However, our research<br />

design recommended an increase only if the investor asked for an increase.<br />

Even these investors had sufficient financial assets to cover this net investment,<br />

as shown by the ratio of the recommended portfolio divided by total financial<br />

wealth (median 48%; even the ninety-fifth percentile of the distribution had<br />

enough financial assets to cover the required net investment with a ratio of<br />

97%) in Table 1, panel D. Moreover, we use the ratio of the recommended<br />

portfolio size to the original portfolio size as another independent variable in<br />

Table 8. As can be seen, this variable has no significant effect on the decision<br />

of whether or not to follow advice.<br />

Third, one could argue that the advice was short-lived and therefore ignored.<br />

Again, the evidence suggests otherwise. Our clients trade often; their average<br />

holding period is only a little more than one year. The advice does not<br />

seem short-lived compared with their trading horizons. Moreover, even if the<br />

free advice ceased after a test phase, there should be no utility loss from<br />

investing in an efficient portfolio. Fourth, it is conceivable that investors may<br />

ignore the advice merely because they are not sophisticated. However, our<br />

results show that subjects who choose to receive the advice are actually more<br />

16 As pensions in Germany are provided by the state and the employer, such a concentration of savings in one<br />

brokerage does not seem excessive.<br />

1004<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

sophisticated, as measured by their past investment performance, than are their<br />

peers who do not accept the offer. Fifth, investors may also choose to ignore<br />

the advice because they do not trust the brokerage, but our results also show<br />

that customers who opt for the advice have had a long relationship with the<br />

brokerage (on average 9.1 years), even significantly longer than non-advisees.<br />

Sixth, is inertia causing them to ignore the advice? Subjects do not seem<br />

to be inert because they trade actively after the recommendation was given,<br />

though they do not follow the recommendation. Seventh, it is also unlikely<br />

that the investors are simply too busy to heed the advice, considering that they<br />

trade actively. Eighth, it could be that clients follow advice only after a certain<br />

time. We recomputed our dependent variable in Table 8—degree of following<br />

from t = 0 to t = 20—for various other time durations. Our results do not<br />

qualitatively change. Ninth, it could be that clients do not buy the stocks that<br />

they are unfamiliar with. We defined a stock to be familiar if an advisee had<br />

traded it in the past and then computed the proportion of stocks in the recommended<br />

portfolio that an advisee was familiar with. This variable also had no<br />

effect in Table 8. Tenth, advisees may not heed advice because they are too<br />

risk-averse. Though the recommendations were tailored for their risk aversion,<br />

we checked whether this variable had an effect in Table 8. The answer is no.<br />

To summarize, our results cannot pinpoint why financial advice is ignored<br />

after it is sought. The only results we have are that wealthier investors and<br />

investors with lower risky portfolio values tend to follow the advice more. We<br />

suspect the lack of results, with respect to the other variables in this test, is due<br />

to a lack of power; there is little variation in the dependent variable. As seen<br />

in Figure 2, panel C, most clients who opt for advice do not follow the advice.<br />

The other reason could be that a systematic cause may not exist to explain why<br />

people do not follow advice.<br />

6. Does the Advice Benefit the Advisee?<br />

We now come to the most critical issue addressed by our study—whether advisees<br />

benefit from financial advice. Many previous studies have documented<br />

that financial advice does not benefit the advisee, but the reason that is given as<br />

the root cause of the finding is that the advice is conflicted. Our research design<br />

ensured that the financial advice was unbiased, un-conflicted, and theoretically<br />

sound ex ante. Was this true ex post, too?<br />

6.1 Is the financial advice sound?<br />

From the description of the data set and the timeline, it follows that the natural<br />

research design to investigate the quality of financial advice is to examine its<br />

effect by using a difference-in-difference methodology. This requires calculating<br />

the improvement of portfolio efficiency between the pre-advice period<br />

and the post-advice period for the treatment group (clients who opt to receive<br />

advice) if they had fully followed the advice and comparing this difference<br />

1005<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

with the improvement of portfolio efficiency between the pre-advice period<br />

and the post-advice period for the control group (clients who opt not to receive<br />

advice). To do this, we define a recommended portfolio as the portfolio that the<br />

treatment group would hold had they completely followed the advice given.<br />

We use two measures of diversification—HHI and the share of idiosyncratic<br />

risk (the part of the risk that is uncompensated)—and two measures of portfolio<br />

performance—the Sharpe ratio and MPPM—as our four measures of portfolio<br />

efficiency. The HHI for the pre-advice period is computed from the actual<br />

portfolios of both the treatment group and the control group at t = 0. The<br />

HHI for the post-advice period for the treatment group is computed on the<br />

recommended portfolios of the treatment group at t = 11. The HHI for the postadvice<br />

period for the control group is computed from their actual portfolios<br />

at t = 11. The share of idiosyncratic variance in the pre-advice period is<br />

computed by running the regression in Equation (1) on the actual portfolios of<br />

both the treatment group and the control group in the period from September<br />

2005 to May 2009. The share of idiosyncratic variance in the post-advice<br />

period for the treatment group is computed by running the regression in<br />

Equation (1) on the recommended portfolios of the treatment group in the<br />

period from October 2009 to April 2010. This computation is possible because<br />

the date and the details of each of the recommendations are known. The share<br />

of idiosyncratic risk in the post-advice period for the control group is computed<br />

by running the regression in Equation (1) on the actual portfolios of the control<br />

group in the period from October 2009 to April 2010. Sharpe ratios and MPPM<br />

in the pre-advice period are computed for the actual portfolios of both the<br />

treatment group and the control group in the period from September 2005 to<br />

May 2009. Sharpe ratios and MPPM in the post-advice period for the treatment<br />

group are computed for the recommended portfolios of the treatment group in<br />

the period from October 2009 to April 2010. Sharpe ratios and MPPM in the<br />

post-advice period for the control group are computed for the actual portfolios<br />

of the control group in the period from October 2009 to April 2010.<br />

If the decrease (improvement) in the HHI or share of idiosyncratic variance<br />

from the pre-advice period to the post-advice period for the treatment group<br />

is greater than those of the control group, it would suggest that the advice is<br />

theoretically sound. If the increase (improvement) in the Sharpe ratio or MPPM<br />

from the pre-advice period to the post-advice period for the treatment group is<br />

greater than those of the control group, it would suggest that the advice is<br />

theoretically sound.<br />

We run four sets of OLS regressions. The dependent variable in our first<br />

set of OLS regressions is the decrease in HHI from the pre-advice period<br />

to the post-advice period. The dependent variable in our second set of OLS<br />

regressions is the decrease in the share of idiosyncratic variance from the preadvice<br />

period to the post-advice period. The dependent variable in our third<br />

set of OLS regressions is the increase in Sharpe ratio from the pre-advice<br />

period to the post-advice period. The dependent variable in our fourth set<br />

1006<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

of OLS regressions is the increase in MPPM from the pre-advice period to<br />

the post-advice period. The main independent variable of interest in all the<br />

four sets of OLS regressions is a “dummy advice” variable set to one for the<br />

treatment group and zero for the control group. Table 9 reports the results of<br />

these regressions.<br />

The coefficient on the “dummy advice” variable is positive and statistically<br />

significant in all four sets of regressions. This implies that if advisees had<br />

fully followed the advice, they would have improved their portfolio efficiency<br />

(decreased the HHI and share of idiosyncratic risk in their portfolios and<br />

increased the Sharpe ratios and MPPM of their portfolios). Thus, the advice<br />

does improve portfolio efficiency.<br />

6.2 Does advice benefit the average advisee?<br />

We use the same research design as described in the previous subsection, with<br />

one important change. Instead of computing the HHI, share of idiosyncratic<br />

risk, Sharpe ratio, and MPPM in the post-advice period for the treatment group<br />

by using their recommended portfolios, we instead use their actual portfolios.<br />

The coefficient on the “dummy advice” variable is statistically insignificant<br />

in the HHI and the idiosyncratic risk share cases but is positively significant<br />

in the Sharpe ratio case and is positively significant in one out of four<br />

models in the MPPM case. Comparing the coefficients in these latter significant<br />

cases with their counterparts in Table 8 reveals that the coefficients are<br />

much smaller (of the order of one-third to one-half). All this evidence implies<br />

that the average advisee does not improve portfolio diversification from the<br />

advice but does improve portfolio performance, though very modestly.<br />

Why does the average advisee not much benefit? Given that the advice<br />

was theoretically sound, it follows that the average advisee did not much<br />

benefit because he or she did not follow the advice. We provided evidence<br />

that the average advisee does not follow the advice in subsection 3.2. We now<br />

investigate whether advice benefits the advisee if he or she partially follows<br />

the advice. To do this, we run the same regressions we ran for Table 10, with<br />

an additional independent variable: the variable measures the increase in the<br />

degree of following the advice between the time a client first received the<br />

advice and the end of the post-advice period (see Figure 2, panel C). Recall<br />

that this variable is positive (negative) if the advisee is partially or fully acting<br />

on (acting against) the advice. Table 11 shows the results.<br />

The coefficient on the “increase in the degree of following” measure is<br />

statistically significant if we use HHI as a diversification measure but is<br />

statistically insignificant if we use the idiosyncratic risk share as our measure<br />

of diversification. The coefficient on the “increase in the degree of following”<br />

measure is statistically significant if we use Sharpe ratio as a portfolio<br />

performance measure but is statistically insignificant if we use MPPM as<br />

a portfolio performance measure. Our results thus show that even partially<br />

following the advice would have improved the efficiency of the portfolio.<br />

1007<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table 9<br />

Does financial advice improve diversification and portfolio performance?<br />

Decrease in HHI Decrease in idiosyncratic variance share<br />

1008<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7) (8)<br />

Dummy advice 0.088∗∗∗ 0.091∗∗∗ 0.091∗∗∗ 0.091∗∗∗ 0.082∗∗∗ 0.087∗∗∗ 0.086∗∗∗ 0.086∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Dummy male −0.000 −0.000 −0.000 0.007 0.005 0.004<br />

(0.993) (0.983) (0.959) (0.309) (0.523) (0.549)<br />

Age −0.000 −0.000 −0.000 0.000 0.000 0.000<br />

(1.000) (0.948) (0.871) (0.687) (0.765) (0.796)<br />

Dummy low wealth 0.003 0.004 0.004 −0.006 −0.008 −0.008<br />

(0.714) (0.691) (0.693) (0.618) (0.530) (0.525)<br />

Dummy high wealth 0.003 0.003 0.002 0.008 0.009 0.009<br />

(0.430) (0.499) (0.558) (0.144) (0.116) (0.115)<br />

Log portfolio value (t = −44) −0.001 0.002 0.002 −0.018∗∗∗ −0.020∗∗∗ −0.019∗∗∗ (0.698) (0.266) (0.292) (0.000) (0.000) (0.000)<br />

Length of relationship 0.000 0.000 −0.001 −0.001<br />

(0.673) (0.753) (0.257) (0.257)<br />

Trades per month (t = −44 to t = 0) −0.000 −0.000 0.003∗∗ 0.003∗∗ (0.918) (0.938) (0.042) (0.038)<br />

Portfolio turnover (t = −44 to t = 0) −0.007 −0.002 0.113 0.098<br />

(0.924) (0.977) (0.103) (0.155)<br />

Disposition effect (t = −44 to t = 0) −0.009 −0.009 0.012 0.014<br />

(0.282) (0.287) (0.228) (0.165)<br />

Risky share (t = 0) −0.032∗∗∗ −0.032∗∗∗ 0.038∗∗∗ 0.039∗∗∗ (0.000) (0.000) (0.000) (0.000)<br />

Share of tax-free assets (t = 0) −0.020 −0.019 −0.046∗∗∗ −0.045∗∗ (0.266) (0.301) (0.009) (0.012)<br />

Alpha (t = −44 to t = 0) 0.167 −1.549<br />

(0.901) (0.158)<br />

Constant 0.002 0.006 0.018 0.020 0.072∗∗∗ 0.241∗∗∗ 0.274∗∗∗ 0.267∗∗∗ (0.225) (0.732) (0.437) (0.400) (0.000) (0.000) (0.000) (0.000)<br />

Observations 7,255 5,426 5,403 5,402 5,453 5,450 5,444 5,444<br />

R2 0.016 0.021 0.027 0.027 0.009 0.028 0.039 0.040<br />

(continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 9<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Dummy advice 0.193∗∗∗ 0.193∗∗∗ 0.189∗∗∗ 0.190∗∗∗ 0.053∗∗∗ 0.057∗∗∗ 0.041∗∗ 0.083∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.003) (0.002) (0.016) (0.000)<br />

Dummy male −0.020∗∗∗ −0.015∗∗∗ −0.015∗∗∗ 0.015 0.002 −0.016<br />

(0.000) (0.006) (0.004) (0.392) (0.908) (0.317)<br />

Age −0.001∗∗∗ −0.000∗ −0.000∗ −0.001 −0.001 −0.001∗∗ (0.004) (0.075) (0.068) (0.245) (0.209) (0.050)<br />

Dummy low wealth 0.013 0.011 0.011 −0.006 −0.001 −0.008<br />

(0.181) (0.246) (0.260) (0.836) (0.962) (0.745)<br />

Dummy high wealth 0.003 0.004 0.004 0.023 0.024 0.028<br />

(0.456) (0.320) (0.276) (0.352) (0.306) (0.197)<br />

Log portfolio value (t = −44) 0.010∗∗∗ 0.003∗ 0.004∗∗ −0.008 −0.018 0.003<br />

(0.000) (0.062) (0.021) (0.460) (0.161) (0.839)<br />

Length of relationship −0.001 −0.001 0.001 0.002<br />

(0.161) (0.195) (0.816) (0.535)<br />

Trades per month (t = −44 to t = 0) 0.011∗∗∗ 0.011∗∗∗ 0.017 0.020∗∗<br />

(0.000) (0.000) (0.147) (0.022)<br />

Portfolio turnover (t = −44 to t = 0) −0.513∗∗∗ −0.548∗∗∗ 1.108∗∗ 0.034<br />

(0.000) (0.000) (0.029) (0.906)<br />

Disposition effect (t = −44 to t = 0) 0.008 0.012 −0.232∗∗∗ −0.110∗∗∗ (0.307) (0.131) (0.001) (0.000)<br />

Risky share (t = 0) 0.008 0.009 0.085 0.107<br />

(0.252) (0.169) (0.215) (0.107)<br />

(continued)<br />

1009<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1010<br />

Table 9<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Share of tax-free assets (t = 0) 0.064∗∗∗ 0.065∗∗∗ 0.072 0.154∗∗ (0.000) (0.000) (0.396) (0.038)<br />

Alpha (t = −44 to t = 0) −3.510∗∗∗ −93.799∗∗∗ (0.000) (0.000)<br />

Constant 0.199∗∗∗ 0.138∗ ∗ ∗ 0.146∗∗∗ 0.131∗∗∗ 0.330∗∗∗ 0.429∗ ∗ ∗ 0.365∗∗∗ −0.036<br />

(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.003) (0.804)<br />

Observations 5,479 5,476 5,470 5,469 5,487 5,484 5,478 5,477<br />

R2 0.079 0.091 0.141 0.148 0.000 0.001 0.019 0.204<br />

Table 9 reports OLS estimates of the coefficients related to a decrease in HHI (models 1 to 4), a decrease in idiosyncratic variance share (models 5 to 8), an increase in Sharpe ratio<br />

(models 9 to 12), and an increase in MPPM (models 13 to 16). HHI and portfolio returns are calculated by using the assumption that investors had fully followed the recommendations<br />

(recommended portfolios). The focus of the table is on the variable Dummy advice that is equal to one if a client opts for financial advice. Additionally, the model controls for several other<br />

independent variables: a dummy that is equal to one if a client is male (Dummy male), the age of a client (Age), a dummy that is equal to one if a client falls into categories one to three of a<br />

microgeographic status rating by an external agency (Dummy low wealth), a dummy that is equal to one if a client falls into categories seven to nine of the microgeographic status (Dummy<br />

high wealth), the risky portfolio value of the customer (Log portfolio value), the number of years the client has been with the bank (Length of relationship), the number of trades per month<br />

(Trades per month), the average portfolio turnover per month (Portfolio turnover), the difference between the proportion of realized gains and losses (Disposition effect), the proportion of<br />

risky assets in the account (Risky share), the proportion of tax-free assets (Share of tax-free assets), and the weekly alpha of a particular customer before opting for financial advice (Alpha).<br />

Alpha and idiosyncratic variance share stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number<br />

of observations are reported. *** denotes significance at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used.<br />

Different counts of observations are due to data availability of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 10<br />

Does the average advisee benefit?<br />

Decrease in HHI Decrease in idiosyncratic variance share<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7) (8)<br />

Dummy advice 0.013 0.011 0.010 0.011 −0.003 0.002 0.001 0.002<br />

(0.144) (0.298) (0.321) (0.304) (0.807) (0.886) (0.907) (0.858)<br />

Dummy male −0.001 −0.000 −0.000 0.005 0.003 0.003<br />

(0.901) (0.961) (0.940) (0.475) (0.637) (0.667)<br />

Age −0.000 −0.000 −0.000 0.000 0.000 0.000<br />

(0.598) (0.619) (0.553) (0.635) (0.647) (0.676)<br />

Dummy low wealth 0.003 0.004 0.004 −0.005 −0.006 −0.006<br />

(0.713) (0.689) (0.691) (0.713) (0.608) (0.603)<br />

Dummy high wealth 0.002 0.001 0.001 0.007 0.008 0.008<br />

(0.681) (0.743) (0.814) (0.189) (0.153) (0.151)<br />

Log portfolio value (t = −44) 0.000 0.003 0.003 −0.017∗∗∗ −0.019∗∗∗ −0.019∗∗∗ (0.899) (0.131) (0.151) (0.000) (0.000) (0.000)<br />

Length of relationship 0.000 0.000 −0.002 −0.002<br />

(0.756) (0.839) (0.126) (0.126)<br />

Trades per month (t = −44 to t = 0) 0.001 0.001 0.003∗∗ 0.003∗∗ (0.390) (0.376) (0.017) (0.015)<br />

Portfolio turnover (t = −44 to t = 0) −0.050 −0.045 0.077 0.062<br />

(0.464) (0.519) (0.266) (0.371)<br />

Disposition effect (t = −44 to t = 0) −0.009 −0.009 0.016 0.017∗ (0.284) (0.283) (0.132) (0.089)<br />

Risky share (t = 0) −0.036∗∗∗ −0.036∗∗∗ 0.035∗∗∗ 0.036∗∗∗ (0.000) (0.000) (0.000) (0.000)<br />

Share of tax-free assets (t = 0) −0.017 −0.015 −0.037∗∗ −0.035∗∗ (0.367) (0.407) (0.039) (0.048)<br />

Alpha (t = −44 to t = 0) 0.243 −1.611<br />

(0.857) (0.144)<br />

Constant 0.002 0.003 0.017 0.019 0.072∗∗∗ 0.232∗∗∗ 0.265∗∗∗ 0.258∗∗∗ (0.225) (0.853) (0.479) (0.432) (0.000) (0.000) (0.000) (0.000)<br />

Observations 7,251 5,422 5,399 5,398 5,453 5,450 5,444 5,444<br />

R2 0.000 0.000 0.007 0.007 0.000 0.017 0.026 0.027<br />

(continued)<br />

1011<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1012<br />

Table 10<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Dummy advice 0.060∗∗∗ 0.060∗∗∗ 0.056∗∗∗ 0.057∗∗∗ 0.010 0.014 −0.001 0.040∗∗ (0.000) (0.000) (0.000) (0.000) (0.614) (0.473) (0.938) (0.027)<br />

Dummy male −0.021∗∗∗ −0.015∗∗∗ −0.016∗∗∗ 0.015 0.002 −0.016<br />

(0.000) (0.005) (0.003) (0.384) (0.899) (0.326)<br />

Age −0.001∗∗∗ −0.000∗ −0.000∗ −0.001 −0.001 −0.001∗∗ (0.002) (0.056) (0.051) (0.231) (0.198) (0.046)<br />

Dummy low wealth 0.011 0.009 0.008 −0.007 −0.003 −0.010<br />

(0.288) (0.377) (0.395) (0.788) (0.914) (0.699)<br />

Dummy high wealth 0.003 0.004 0.004 0.022 0.024 0.027<br />

(0.513) (0.364) (0.317) (0.365) (0.318) (0.207)<br />

Log portfolio value (t = −44) 0.010∗∗∗ 0.004∗∗ 0.005∗∗ −0.008 −0.018 0.003<br />

(0.000) (0.042) (0.014) (0.449) (0.161) (0.841)<br />

Length of relationship −0.001 −0.001 0.001 0.002<br />

(0.139) (0.169) (0.815) (0.539)<br />

Trades per month (t = −44 to t = 0) 0.011∗∗∗ 0.011∗∗∗ 0.018 0.020∗∗ (0.000) (0.000) (0.143) (0.021)<br />

Portfolio turnover (t = −44 to t = 0) −0.520∗∗∗ −0.555∗∗∗ 1.116∗∗ 0.043<br />

(0.000) (0.000) (0.028) (0.882)<br />

Disposition effect (t = −44 to t = 0) 0.008 0.012 −0.233∗∗∗ −0.112∗∗∗ (0.322) (0.144) (0.001) (0.000)<br />

Risky share (t = 0) 0.004 0.006 0.082 0.104<br />

(0.537) (0.401) (0.234) (0.119)<br />

(continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 10<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Share of tax-free assets (t = 0) 0.063∗∗∗ 0.065∗∗∗ 0.074 0.156∗∗ (0.000) (0.000) (0.381) (0.035)<br />

Alpha (t = −44 to t = 0) −3.417∗∗∗ −93.756∗∗∗ (0.000) (0.000)<br />

Constant 0.199∗∗∗ 0.138∗∗∗ 0.148∗∗∗ 0.133∗∗∗ 0.330∗∗∗ 0.433∗∗∗ 0.367∗∗∗ −0.034<br />

(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.003) (0.813)<br />

Observations 5,479 5,476 5,470 5,469 5,487 5,484 5,478 5,477<br />

R2 0.008 0.021 0.074 0.081 0.000 0.001 0.019 0.203<br />

Table 10 reports OLS estimates of the coefficients related to a decrease in HHI (models 1 to 4) and a decrease in idiosyncratic variance share (models 5 to 8), an increase in Sharpe ratio<br />

(models 9 to 12) and an increase in MPPM (models 13 to 16). HHI and portfolio returns are calculated based on the actual portfolios of investors. The focus of the table is on the variable<br />

Dummy advice that is equal to one if a client opts for financial advice. Additionally, the model controls for the following independent variables: a dummy variable that is equal to one if a<br />

client is male (Dummy male), the age of a client (Age), a dummy that is equal to one if a client falls into categories one to three of a microgeographic status rating by an external agency<br />

(Dummy low wealth), a dummy that is equal to one if a client falls into categories seven to nine of the microgeographic status (Dummy high wealth), the risky portfolio value of the customer<br />

(Log portfolio value), the number of years the client has been with the bank (Length of relationship), the number of trades per month (Trades per month), the average portfolio turnover per<br />

month (Portfolio turnover), the difference between the proportion of realized gains and losses (Disposition effect), the proportion of risky assets in the account (Risky share), the proportion<br />

of tax-free assets (Share of tax-free assets), and the weekly alpha of a particular customer before opting for financial advice (Alpha). Alpha and idiosyncratic variance share stem from the<br />

application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number of observations are reported. *** denotes significance<br />

at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used. Different counts of observations are due to data availability<br />

of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

1013<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table 11<br />

Does the average advisee benefit if the advice is partially followed?<br />

Decrease in HHI Decrease in idiosyncratic variance share<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7) (8)<br />

Improvement of degree of following 0.136∗∗∗ 0.146∗∗∗ 0.147∗∗∗ 0.147∗∗∗ 0.113 0.103 0.119 0.118<br />

(0.002) (0.003) (0.003) (0.004) (0.122) (0.150) (0.100) (0.104)<br />

Dummy advice 0.012 0.009 0.008 0.009 −0.004 0.001 0.000 0.001<br />

(0.177) (0.392) (0.422) (0.403) (0.727) (0.964) (0.998) (0.947)<br />

Dummy male −0.001 −0.000 −0.000 0.005 0.003 0.003<br />

(0.906) (0.960) (0.939) (0.479) (0.645) (0.675)<br />

Age −0.000 −0.000 −0.000 0.000 0.000 0.000<br />

(0.625) (0.647) (0.580) (0.619) (0.628) (0.657)<br />

Dummy low wealth 0.003 0.004 0.004 −0.004 −0.006 −0.006<br />

(0.699) (0.676) (0.678) (0.720) (0.616) (0.610)<br />

Dummy high wealth 0.002 0.001 0.001 0.007 0.008 0.008<br />

(0.709) (0.771) (0.843) (0.195) (0.158) (0.157)<br />

Log portfolio value (t = −44) 0.000 0.003 0.003 −0.017∗∗∗ −0.019∗∗∗ −0.019∗∗∗ (0.876) (0.126) (0.146) (0.000) (0.000) (0.000)<br />

Length of relationship 0.000 0.000 −0.002 −0.002<br />

(0.772) (0.855) (0.121) (0.121)<br />

Trades per month (t = −44 to t = 0) 0.001 0.001 0.003∗∗ 0.003∗∗ (0.354) (0.341) (0.016) (0.014)<br />

Portfolio turnover (t = −44 to t = 0) −0.050 −0.045 0.077 0.062<br />

(0.463) (0.519) (0.266) (0.370)<br />

Disposition effect (t = −44 to t = 0) −0.009 −0.009 0.015 0.017∗<br />

(0.272) (0.270) (0.137) (0.093)<br />

Risky share (t = 0) −0.036∗∗∗ −0.036∗∗∗ 0.036∗∗∗ 0.036∗∗∗ (0.000) (0.000) (0.000) (0.000)<br />

Share of tax-free assets (t = 0) −0.017 −0.016 −0.037∗∗ −0.036∗∗ (0.342) (0.380) (0.036) (0.044)<br />

Alpha (t = −44 to t = 0) 0.259 −1.603<br />

(0.848) (0.147)<br />

Constant 0.002 0.002 0.017 0.019 0.072∗∗∗ 0.232∗∗∗ 0.265∗∗∗ 0.258∗∗∗ (0.225) (0.890) (0.478) (0.430) (0.000) (0.000) (0.000) (0.000)<br />

Observations 7,251 5,422 5,399 5,398 5,453 5,450 5,444 5,444<br />

R2 0.001 0.002 0.008 0.009 0.001 0.017 0.027 0.028<br />

(continued)<br />

1014<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table 11<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Improvement of degree of following 0.193∗∗∗ 0.197∗∗∗ 0.187∗∗∗ 0.185∗∗∗ 0.018 0.006 0.085 0.026<br />

(0.001) (0.001) (0.001) (0.002) (0.752) (0.913) (0.199) (0.645)<br />

Dummy advice 0.058∗∗∗ 0.058∗∗∗ 0.054∗∗∗ 0.055∗∗∗ 0.010 0.014 −0.002 0.040∗∗ (0.000) (0.000) (0.000) (0.000) (0.622) (0.477) (0.898) (0.030)<br />

Dummy male −0.021∗∗∗ −0.015∗∗∗ −0.016∗∗∗ 0.015 0.002 −0.016<br />

(0.000) (0.004) (0.003) (0.384) (0.902) (0.325)<br />

Age −0.001∗∗∗ −0.000∗ −0.000∗ −0.001 −0.001 −0.001∗∗ (0.003) (0.063) (0.058) (0.232) (0.200) (0.047)<br />

Dummy low wealth 0.011 0.009 0.009 −0.007 −0.003 −0.010<br />

(0.278) (0.365) (0.383) (0.789) (0.917) (0.700)<br />

Dummy high wealth 0.003 0.004 0.004 0.022 0.024 0.027<br />

(0.539) (0.383) (0.334) (0.365) (0.320) (0.207)<br />

Log portfolio value (t = −44) 0.010∗∗∗ 0.004∗∗ 0.005∗∗ −0.008 −0.018 0.003<br />

(0.000) (0.039) (0.013) (0.449) (0.161) (0.840)<br />

Length of relationship −0.001 −0.001 0.001 0.002<br />

(0.128) (0.155) (0.818) (0.540)<br />

Trades per month (t = −44 to t = 0) 0.011∗∗∗ 0.011∗∗∗ 0.018 0.020∗∗ (0.000) (0.000) (0.142) (0.021)<br />

Portfolio turnover (t = −44 to t = 0) −0.520∗∗∗ −0.555∗∗∗ 1.116∗∗ 0.043<br />

(0.000) (0.000) (0.028) (0.882)<br />

Disposition effect (t = −44 to t = 0) 0.008 0.012 −0.233∗∗∗ −0.112∗∗∗ (0.342) (0.156) (0.001) (0.000)<br />

Risky share (t =0) 0.004 0.006 0.082 0.104<br />

(0.522) (0.389) (0.234) (0.118)<br />

Share of tax-free assets (t = 0) 0.062∗∗∗ 0.064∗∗∗ 0.074 0.156∗∗ (0.000) (0.000) (0.384) (0.036)<br />

(continued)<br />

1015<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1016<br />

Table 11<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Alpha (t = −44 to t = 0) −3.405∗∗∗ −93.754∗∗∗ (0.000) (0.000)<br />

Constant 0.199∗∗∗ 0.137∗∗∗ 0.148∗∗∗ 0.133∗∗∗ 0.330∗∗∗ 0.433∗∗∗ 0.367∗∗∗ −0.034<br />

(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.003) (0.813)<br />

Observations 5,479 5,476 5,470 5,469 5,487 5,484 5,478 5,477<br />

R2 0.011 0.023 0.077 0.084 0.000 0.001 0.019 0.203<br />

Table 11 reports OLS estimates of the coefficients related to a decrease in HHI (models 1 to 4) and a decrease in idiosyncratic variance share (models 5 to 8), an increase in Sharpe ratio<br />

(models 9 to 12) and an increase in MPPM (models 13 to 16). HHI and portfolio returns are calculated based on the actual portfolios of investors. The focus of the table is on the variable<br />

Improvement of degree of following, which shows the improvement of the degree of following from the time of the first recommendation to t = 11 and takes the value zero for non-advised<br />

clients. Dummy advice is equal to one if a client opts for financial advice. Additionally, the model controls for the following independent variables: a dummy variable that is equal to one if<br />

a client is male (Dummy male), the age of a client (Age), a dummy that is equal to one if a client falls into categories one to three of a microgeographic status rating by an external agency<br />

(Dummy low wealth), a dummy that is equal to one if a client falls into categories seven to nine of the microgeographic status (Dummy high wealth), the risky portfolio value of the customer<br />

(Log portfolio value), the number of years the client has been with the bank (Length of relationship), the number of trades per month (Trades per month), the average portfolio turnover per<br />

month (Portfolio turnover), the difference between the proportion of realized gains and losses (Disposition effect), the proportion of risky assets in the account (Risky share), the proportion<br />

of tax-free assets (Share of tax-free assets), and the weekly alpha of a particular customer before opting for financial advice (Alpha). Alpha and idiosyncratic variance share stem from the<br />

application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number of observations are reported. *** denotes significance<br />

at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used. Different counts of observations are due to data availability<br />

of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

7. Who Would Most Benefit from the Advice?<br />

We now explore an important public policy question: are the individuals who<br />

are most likely to benefit from financial advice the ones least likely to obtain<br />

the advice, and are the persons least likely to benefit from financial advice most<br />

likely to obtain and follow the advice? Section 4 indicates that this seems to<br />

be the case; the evidence showed that the more (less) financially sophisticated<br />

investors are, the more (less) likely they are to obtain advice.<br />

We revisit the probit regression that was run in order to generate the results<br />

in Table 7, i.e., the test that investigates those who elect to receive financial<br />

advice. We take probit model 5 in Table 7 because it has the highest pseudo-<br />

R 2 value and can be estimated for 7,432 customers. We then use the coefficient<br />

estimates from this regression to predict the 5% of the 7,432 clients with<br />

the highest probability of obtaining the advice. 17 We define those 5% (372<br />

investors) as “predicted to obtain advice” and the remaining 95% (7,060<br />

investors) as “not predicted to obtain advice.” Of the former group, sixty-two<br />

opted for advice and 310 did not. Of the latter group, 307 opted for advice and<br />

6,753 did not. Thus, we can test whether those persons predicted to be less<br />

likely to obtain the advice are the ones who benefit more from the advice (i.e.,<br />

whether the 307 benefited more than the sixty-two).<br />

The number of observations in these subgroups is lower than 372 and 7,060<br />

because of data availability. We now run the same regression that generated<br />

the results in Table 9 for both these subgroups. The first subgroup consists of<br />

clients who are predicted to obtain advice. Group size is 372. Of these, sixtytwo<br />

took advice and have the indicator variable “dummy advice” turned on to<br />

equal one, and 310 did not take advice and have the indicator variable “dummy<br />

advice” turned on to equal zero. The second subgroup is clients who are not<br />

predicted to obtain advice. The group size is 7,060. Of these, 307 took advice<br />

and have the indicator variable “dummy advice” turned on to equal one, and<br />

6,753 did not take advice and have the indicator variable “dummy advice”<br />

turned on to equal zero. Table 12 reports the results for these regressions.<br />

The coefficient on the “dummy advice” variable is 0.042 (positive and<br />

statistically significant) in the HHI regression for the clients who are predicted<br />

to opt for the advice and did opt for it. The coefficient on the “dummy advice”<br />

variable is 0.099 (positive and statistically significant) in the HHI regression<br />

for the clients who are predicted to not obtain the advice but did obtain<br />

it. The difference between these two coefficients is positive and statistically<br />

significant.<br />

The coefficient on the “dummy advice” variable is 0.066 (positive and<br />

statistically significant) in the idiosyncratic risk regression for the clients who<br />

are predicted to opt for the advice and did opt for it. The coefficient on the<br />

“dummy advice” variable is 0.077 (positive and statistically significant) in the<br />

17 Results are robust to different cutoff points, including 10%, 20%, or even 50%, with highest likelihood to opt<br />

for advice.<br />

1017<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1018<br />

Table 12<br />

Who would benefit most from advice?<br />

Decrease in HHI Decrease in idiosyncratic variance share<br />

Predicted to Not predicted to Predicted to Not predicted to<br />

Dependent variable obtain advice obtain advice P-value obtain advice obtain advice P-value<br />

Dummy advice 0.042∗∗∗ 0.099∗∗∗ .00 0.066∗∗ 0.077∗∗∗ .73<br />

(0.000) (0.000) (0.013) (0.000)<br />

Constant −0.007∗ 0.003 0.112∗∗∗ 0.070∗∗∗ (0.089) (0.135) (0.000) (0.000)<br />

Observations 366 6,859 359 5,085<br />

R2 0.056 0.017 0.018 0.007<br />

Increase in Sharpe ratio Increase in MPPM<br />

Predicted to Not predicted to Predicted to Not predicted to<br />

Dependent variable obtain advice obtain advice P-value obtain advice obtain advice P-value<br />

Dummy advice 0.147∗∗∗ 0.194∗∗∗ .00 −0.034 0.053∗∗∗ .18<br />

(0.000) (0.000) (0.579) (0.006)<br />

Constant 0.250∗∗∗ 0.196∗∗∗ 0.439∗∗∗ 0.324∗∗∗ (0.000) (0.000) (0.000) (0.000)<br />

Observations 359 5,110 359 5,118<br />

R2 0.179 0.069 0.000 0.000<br />

Table 12 reports coefficient estimates of the variable “dummy advice” when the tests of Table 9 are conducted on subgroups. The first subgroup consists of clients who are predicted to<br />

obtain advice. These are 5% of the 7,432 clients with the highest probability of obtaining the advice predictions as given in model 5 of Table 7. Their size is 372. Of these, sixty-two took<br />

advice and so have the indicator variable “dummy advice” turned on to equal one, and 310 did not take advice and so have the indicator variable “dummy advice” turned on to equal zero.<br />

The second subgroup consists of clients who are not predicted to obtain advice. These are 95% of the 7,432 clients with the lowest probability of obtaining the advice predictions as given<br />

in model 5 of Table 7. Their size is 7,060. Of these, 307 took advice and so have the indicator variable “dummy advice” turned on to equal one, and 6,753 did not take advice and so have<br />

the indicator variable “dummy advice” turned on to equal zero. R2 values and number of observations are reported. *** denotes significance at 1% or less, ** significance at 5% or less,<br />

and * significance at 10% or less. Heteroscedasticity robust standard errors are used. The third column reports the p-value of a test of equality of the coefficients for Dummy advice using<br />

seemingly unrelated estimation. Different counts of observations are due to data availability of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

idiosyncratic risk regression for the clients who are predicted to not obtain<br />

the advice but did obtain it. The difference between these two coefficients is<br />

positive but not statistically significant.<br />

The coefficient on the “dummy advice” variable is 0.147 (positive and<br />

statistically significant) in the Sharpe ratio regression for the clients who are<br />

predicted to opt for the advice and did opt for it. The coefficient on the “dummy<br />

advice” variable is 0.194 (positive and statistically significant) in the Sharpe<br />

ratio regression for the clients who are predicted to not obtain the advice but<br />

did obtain it. The difference between these two coefficients is positive and<br />

statistically significant.<br />

The coefficient on the “dummy advice” variable is -0.034 (negative but<br />

not statistically significant) in the MPPM regression for the clients who are<br />

predicted to opt for the advice and did opt for it. The coefficient on the “dummy<br />

advice” variable is 0.053 (positive and statistically significant) in the MPPM<br />

regression for the clients who are predicted to not obtain the advice but did<br />

obtain it. The difference between these two coefficients is positive but not<br />

statistically significant.<br />

The important result here is that the decrease in HHI or the increase in the<br />

Sharpe ratio is significantly greater for clients predicted to not obtain the advice<br />

but who obtained it than for clients who are predicted to obtain the advice and<br />

did obtain it. As seen in Section 4, the former group includes the investors who<br />

are less financially sophisticated than are the latter group. We come to the same<br />

conclusion: those who most need (do not need) the financial advice are least<br />

(most) likely to obtain it.<br />

8. Conclusion<br />

Can unbiased financial advice steer retail investors toward efficient portfolios?<br />

To answer this question, we work with one of the biggest brokerages in<br />

Germany and offer advice that is unbiased and theoretically sound. First, we<br />

find that those who accept the offer (5%) are more likely to be male, older,<br />

wealthier, more financially sophisticated, and have a longer relationship with<br />

the brokerage. Second, of those who accept the offer, the advice is hardly<br />

followed. Third, though the average advisee’s portfolio efficiency hardly<br />

improves, the average advisee who follows the advice does see an increase<br />

in portfolio efficiency. Fourth, it seems that investors who most need the<br />

financial advice are least likely to obtain it. Overall, our results imply that the<br />

mere availability of unbiased financial advice is a necessary but not sufficient<br />

condition for benefiting retail customers. As the adage goes, you can lead a<br />

horse to water, but you can’t make it drink.<br />

At a time when protecting financial consumers has risen to the top of<br />

the regulatory agenda in many countries, the results of this article provide a<br />

basis for skepticism of supply-side solutions imposed by regulators. However,<br />

examining the demand side of financial advice raises more questions than<br />

1019<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

answers. We learn that even honest and sound financial services are useless,<br />

unless the customer actually follows them. In addition, it seems even more<br />

difficult to reach people who most need financial advice. Thus, if financial<br />

economists are to develop remedies to correct widespread investment mistakes<br />

made by retail investors, this study indicates that much additional research<br />

must be done to answer the question of why people do or do not follow<br />

advice. This must be taken into account when trying to help people make<br />

asset allocation decisions. A promising start has been made by Benartzi and<br />

Thaler (2004, 2007) who, after taking behavioral factors into account, develop<br />

sophisticated savings plans.<br />

Experimenting with ways to offer advice is another useful avenue to explore.<br />

For example, in our study, the advice would require people to turn over, on<br />

average, 75% of their portfolios, since investors’ existing portfolios are largely<br />

inefficient. Investors in our sample may have found it too complicated or too<br />

cumbersome to implement the full list of recommendations, though they did<br />

turn over 70% of their portfolios every year during the pre-advice period. Although<br />

the information given to our advisees is extensive and clear, it may not<br />

be much different from other, less theoretically anchored sources of investment<br />

advice. Perhaps future settings could therefore seek to build greater trust with<br />

advisees. 18 To conclude, much more needs to be done to understand why and<br />

how financial advice is actually followed. What makes the horse drink?<br />

Appendix A: Disguised Example of Advice That Was Sent to Advisees<br />

1. Description of the idea of diversification, explanation of important concepts, intuitive<br />

explanation of the portfolio optimization methodology, and discussion of tax implications<br />

2. Analysis of the existing portfolio<br />

3. Analysis of recommended (optimized) portfolio 19<br />

18 Bonaccio and Dalal (2006) conduct a literature review to document the determinants of all effective advice, not<br />

just financial advice. A variety of characteristics of the advice are examined, including the distance of the advice<br />

from the original opinion (Yaniv 2004); whether the advice was paid for (Gino 2008); whether advice is didactic<br />

or just offers information about choices (Bonaccio and Dalal 2010); and many other aspects. Statman (2010)<br />

explores what investors really want.<br />

19 Recall that the goal of the optimizer is to enhance portfolio efficiency by improving diversification. This may<br />

improve portfolios along two dimensions: either the risk for a given level of return decreases or the expected<br />

return for a given level of risk increases. The brokerage communicated in this report changes along both<br />

dimensions.<br />

1020<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Model calculation of possible portfolio development in 5 years<br />

Lower upper<br />

Current value Expected value bound (5%) bound (95%)<br />

Existing portfolio<br />

Recommended (optimized)<br />

1,000 e 1,700 e 1,100 e 2,500 e<br />

portfolio 1,000 e 2,200 e 1,000 e 4,100 e<br />

Model calculation assumes constant riek/return estimates<br />

4. The client’s investment requests<br />

5. List of necessary transactions<br />

6. Fact sheet for each security on recommendation list<br />

Appendix B: Measure for the Degree of Following<br />

Our degree of following measures both the buy and sell sides of following advice. If a security<br />

in the advisee’s portfolio is not included in the recommended portfolio, this suggests that the<br />

investor has been advised to sell this security. It is, however, possible that the advisee sells a<br />

security for reasons other than following the advice (e.g., liquidity or tax motives). As a check for<br />

robustness, we construct an alternative measure of the degree of following. This measure considers<br />

only the buy side and is simply the share of the recommended portfolio that the investor holds at<br />

any time.<br />

The tables below that use this alternative measure of the degree of following (Tables A1–A4)<br />

are the counterparts of the tables in the text (Table 4, Table 6, Table 8, and Table 11).<br />

The figures below that use this alternative measure of the degree of following (Figure A1,<br />

panels A–C) are the counterparts of the figures in the text (Figure 2, panels A–C).<br />

1021<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1022<br />

Table A1<br />

Illustration of the alternative degree of following (buy side only). Counterpart of Table 4<br />

Required Keep Buy<br />

Original portfolio (t = 0) e Recommended portfolio (t = 0) e action e e<br />

⎫<br />

Deutsche Bank 8,646 → Sell<br />

HSBC Indian Equity Fund 3,622 → Sell ⎪⎬<br />

Ignored for<br />

Raiffeisen CEE Equity Fund 2,792 → Sell measuring<br />

HSBC BRIC Equity Fund 1,862 → Sell ⎪⎭ alternative “degree<br />

Commerzbank 439 → Sell of following advice”<br />

⎫<br />

E.ON 2,523 E.ON 681 → Keep/Decrease 681<br />

H&M 3,543 H&M 3,543 → Keep 3,543<br />

Comstage ETF EONIA 4,072 → Buy 4,072<br />

Schroder ISF Europa Corporate Bond Fund 3,883 → Buy 3,883⎪⎬ Considered for<br />

Allianz Pimco Europazins Bond Fund 3,799 → Buy 3,799 measuring<br />

Allianz Pimco Corporate Bond Europa Fund 2,434 → Buy 2,434 alternative “degree<br />

UBS Lux Bond Fund 1,751 → Buy 1,751 of following advice”<br />

Grundbesitz Europa Real Estate Fund 1,470 → Buy 1,470<br />

Pictet EM Fund 1,247 → Buy 1,247⎪⎭<br />

Allianz RCM Small Cap Fund 808 → Buy 808<br />

Total 23,426 23,688 4,224 19,463<br />

Table A1 provides an actual example of our alternative measure of degree of following, which considers only the assets the advisee is recommended to buy. Securities E.ON and H&M<br />

overlap between the actual and the recommended portfolio (E.ON only partially). The value of the overlap in this case is e4,224 (e681 in E.ON and e3,543 in H&M). We calculate the<br />

degree of following for each day as the euro-value of recommended assets in the investor’s original portfolio divided by the value of assets in euros in the actual portfolio (here: 4,224 /<br />

23,688 = 18%).<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table A2<br />

Summary statistics of the alternative degree of following (buy side only). Counterpart of Table 6<br />

t = 11 months Average<br />

(end of between t =<br />

measurement advice start and<br />

t = advice start t = 10 days t = 20 days t = 30 days period) t = 11 months<br />

Degree of following<br />

Mean (%) 25.0 32.7 36.0 37.8 26.3 32.9<br />

Median (%) 21.4 26.3 27.4 28.9 21.0 28.1<br />

Number of followers n/a 99 142 169 148 175<br />

Observations 385 385 385 385 385 385<br />

Table A2 reports summary statistics of the alternative degree of following. The table reports mean alternative degree of following, median alternative degree of following, number of<br />

followers, and observations for six distinct time periods (upon receiving the first recommendation, after ten days, after twenty days, and at the end of the post-advice period and the average<br />

over the entire post-advice period). These time intervals are specific to each investor in the sample. Number of followers is defined as the number of investors who increase their alternative<br />

degree of following, at least marginally.<br />

1023<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table A3<br />

Who chooses to follow advice? (buy side only) Counterpart of Table 8<br />

1024<br />

Increase in alternative degree of following (buy side only)<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Dummy male −0.063 −0.062 −0.041 −0.049 −0.066 −0.065 −0.028<br />

(0.223) (0.226) (0.444) (0.361) (0.197) (0.203) (0.605)<br />

Age 0.000 0.000 0.000 0.000 0.000 0.000 0.000<br />

(0.619) (0.746) (0.896) (0.871) (0.807) (0.825) (0.664)<br />

Dummy low wealth −0.048∗ −0.061∗∗ −0.071∗∗ −0.066∗∗ −0.059∗ −0.058∗ −0.062∗∗ (0.075) (0.046) (0.016) (0.033) (0.051) (0.064) (0.048)<br />

Dummy high wealth 0.046∗ 0.042 0.031 0.031 0.042 0.041 0.030<br />

(0.073) (0.114) (0.266) (0.259) (0.114) (0.121) (0.305)<br />

Log portfolio value (t =0) −0.068∗∗∗ −0.069∗∗∗ −0.065∗∗∗ −0.061∗∗∗ −0.068∗∗∗ −0.070∗∗∗ −0.059∗∗∗ (0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.000)<br />

Length of relationship 0.002 0.004 0.004 0.001 0.002 0.003<br />

(0.685) (0.454) (0.509) (0.786) (0.712) (0.635)<br />

Trades per month (t = −44 to t =0) 0.002 0.002 0.003 0.004 0.003 0.001<br />

(0.639) (0.729) (0.482) (0.472) (0.537) (0.804)<br />

Portfolio turnover (t = −44 to t =0) −0.151 −0.017 −0.169 −0.198 −0.176 −0.116<br />

(0.590) (0.955) (0.569) (0.504) (0.537) (0.800)<br />

Disposition effect (t = −44 to t =0) −0.007 0.016 0.033 −0.006 −0.010 0.062<br />

(0.899) (0.782) (0.563) (0.917) (0.853) (0.288)<br />

Risky share (t =0) −0.016 −0.036 −0.041 −0.021 −0.014 −0.013<br />

(0.764) (0.513) (0.461) (0.712) (0.787) (0.837)<br />

Share of tax-free assets (t =0) 0.008 0.058 0.092 0.017 0.013 0.080<br />

(0.910) (0.310) (0.160) (0.816) (0.854) (0.141)<br />

Recommended portfolio over original portfolio 0.001 0.000 0.000 −0.002 0.000 0.026<br />

(0.951) (0.985) (0.992) (0.904) (0.987) (0.245)<br />

Alpha (t = −44 to t =0) 7.230<br />

(0.161)<br />

Idiosyncratic variance share (t = −44 to t =0) 0.100<br />

(0.202)<br />

HHI (t =0) 0.099<br />

(0.483)<br />

Home bias (t =0) 0.030<br />

(0.470)<br />

(continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table A3<br />

Continued<br />

Increase in alternative degree of following (buy side only)<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7)<br />

Long-term raw return (t = −44 to t = −1) 8.582<br />

(0.208)<br />

Short-term raw return (t = −1 to t = 0) −1.995∗ (0.093)<br />

Constant 0.842∗∗∗ 0.859∗∗∗ 0.759∗∗∗ 0.656∗∗∗ 0.842∗∗∗ 0.859∗∗∗ 0.668∗∗∗ (0.000) (0.000) (0.000) (0.002) (0.000) (0.000) (0.000)<br />

Observations 369 365 313 313 365 365 297<br />

R2 0.118 0.115 0.126 0.127 0.118 0.117 0.123<br />

Table A3 reports OLS estimates of the coefficients related to an increase in the alternative degree of following measure from the date the investor received advice to t = 20 days (Increase<br />

in the alternative degree of following). For the estimation of the model, we include the following independent variables: a dummy that is equal to one if a client is male (Dummy male), the<br />

age of a client (Age), a dummy that is equal to one if a client falls into categories one to three of a microgeographic status rating by an external agency (Dummy low wealth), a dummy that<br />

is equal to one if a client falls into categories seven to nine of the microgeographic status (Dummy high wealth), the risky portfolio value of the customer (Log portfolio value), the number<br />

of years the client has been with the bank (Length of relationship), the number of trades per month (Trades per month), the average portfolio turnover per month (Portfolio turnover), the<br />

difference between the proportion of realized gains and losses (Disposition effect), the proportion of risky assets in the account (Risky share), the proportion of tax-free assets (Share of<br />

tax-free assets), a ratio for the required net investment (Recommended portfolio/original portfolio), the weekly alpha of a particular customer before opting for financial advice (Alpha),<br />

the idiosyncratic variance share (Idiosyncratic variance share), the Herfindahl-Hirschman index (HHI), the share of domestic equity (Home bias), the raw return from the beginning of<br />

observation until one month prior to the offer of financial advice (Long-term raw return), and the raw return of the month before the offer of financial advice (Short-term raw return).<br />

Alpha and idiosyncratic risk share stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number of<br />

observations are reported. *** denotes significance at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used. Standard<br />

errors shown are not clustered, but results remain qualitatively unaltered when clustering them by advice week or risk aversion. Different counts of observations are due to data availability<br />

of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

1025<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Table A4<br />

Does the average advisee benefit if partially followed? (buy side only) Counterpart of Table 11<br />

Decrease in HHI Decrease in idiosyncratic variance share<br />

1026<br />

Dependent variable (1) (2) (3) (4) (5) (6) (7) (8)<br />

Improvement of alternative degree of following 0.176∗∗∗ 0.197∗∗∗ 0.196∗∗∗ 0.196∗∗∗ 0.139∗∗ 0.134∗∗ 0.149∗∗ 0.148∗∗ (0.001) (0.001) (0.001) (0.001) (0.025) (0.029) (0.016) (0.017)<br />

Dummy advice 0.011 0.007 0.006 0.007 −0.005 −0.001 −0.001 −0.000<br />

(0.218) (0.506) (0.543) (0.521) (0.661) (0.965) (0.925) (0.976)<br />

Dummy male −0.001 −0.000 −0.001 0.005 0.003 0.003<br />

(0.874) (0.924) (0.903) (0.492) (0.662) (0.693)<br />

Age −0.000 −0.000 −0.000 0.000 0.000 0.000<br />

(0.626) (0.649) (0.582) (0.621) (0.630) (0.658)<br />

Dummy low wealth 0.004 0.004 0.004 −0.004 −0.006 −0.006<br />

(0.688) (0.666) (0.668) (0.724) (0.619) (0.613)<br />

Dummy high wealth 0.001 0.001 0.001 0.007 0.008 0.008<br />

(0.738) (0.799) (0.872) (0.203) (0.165) (0.164)<br />

Log portfolio value (t = −44) 0.000 0.003 0.003 −0.017∗∗∗ −0.019∗∗∗ −0.019∗∗∗ (0.878) (0.131) (0.152) (0.000) (0.000) (0.000)<br />

Length of relationship 0.000 0.000 −0.002 −0.002<br />

(0.777) (0.861) (0.119) (0.119)<br />

Trades per month (t = −44 to t =0) 0.001 0.001 0.003∗∗ 0.003∗∗ (0.321) (0.310) (0.014) (0.013)<br />

Portfolio turnover (t = −44 to t =0) −0.051 −0.045 0.077 0.061<br />

(0.456) (0.513) (0.269) (0.373)<br />

Disposition effect (t = −44 to t =0) −0.009 −0.009 0.015 0.017∗ (0.268) (0.265) (0.140) (0.095)<br />

Risky share (t =0) −0.035∗∗∗ −0.036∗∗∗ 0.036∗∗∗ 0.036∗∗∗ (0.000) (0.000) (0.000) (0.000)<br />

Share of tax-free assets (t =0) −0.018 −0.017 −0.038∗∗ −0.036∗∗ (0.326) (0.362) (0.034) (0.042)<br />

Alpha (t = −44 to t =0) 0.269 −1.596<br />

(0.842) (0.149)<br />

Constant 0.002 0.003 0.018 0.020 0.072∗∗∗ 0.232∗∗∗ 0.265∗∗∗ 0.258∗∗∗ (0.225) (0.879) (0.456) (0.408) (0.000) (0.000) (0.000) (0.000)<br />

Observations 7,251 5,422 5,399 5,398 5,453 5,450 5,444 5,444<br />

R2 0.003 0.005 0.011 0.011 0.001 0.018 0.028 0.028<br />

(continued)<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Table A4<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Improvement of alternative degree of following 0.131∗∗∗ 0.134∗∗∗ 0.128∗∗ 0.127∗∗ 0.027 0.018 0.088 0.031<br />

(0.007) (0.008) (0.012) (0.014) (0.625) (0.743) (0.145) (0.553)<br />

Dummy advice 0.058∗∗∗ 0.058∗∗∗ 0.054∗∗∗ 0.055∗∗∗ 0.009 0.014 −0.003 0.040∗∗ (0.000) (0.000) (0.000) (0.000) (0.631) (0.484) (0.876) (0.031)<br />

Dummy male −0.021∗∗∗ −0.015∗∗∗ −0.016∗∗∗ 0.015 0.002 −0.016<br />

(0.000) (0.004) (0.003) (0.385) (0.906) (0.324)<br />

Age −0.001∗∗∗ −0.000∗ −0.000∗ −0.001 −0.001 −0.001∗∗ (0.003) (0.059) (0.054) (0.232) (0.199) (0.047)<br />

Dummy low wealth 0.011 0.009 0.008 −0.007 −0.003 −0.010<br />

(0.280) (0.368) (0.385) (0.789) (0.917) (0.700)<br />

Dummy high wealth 0.003 0.004 0.004 0.022 0.024 0.027<br />

(0.545) (0.389) (0.339) (0.365) (0.321) (0.208)<br />

Log portfolio value (t = −44) 0.010∗∗∗ 0.004∗∗ 0.005∗∗ −0.008 −0.018 0.003<br />

(0.000) (0.041) (0.014) (0.449) (0.161) (0.841)<br />

Length of relationship −0.001 −0.001 0.001 0.002<br />

(0.130) (0.159) (0.818) (0.540)<br />

Trades per month (t = −44 to t =0) 0.011∗∗∗ 0.011∗∗∗ 0.018 0.020∗∗ (0.000) (0.000) (0.142) (0.021)<br />

Portfolio turnover (t = −44 to t =0) −0.521∗∗∗ −0.555∗∗∗ 1.116∗∗ 0.043<br />

(0.000) (0.000) (0.028) (0.882)<br />

Disposition effect (t = −44 to t =0) 0.008 0.012 −0.233∗∗∗ −0.112∗∗∗ (0.338) (0.154) (0.001) (0.000)<br />

Risky share (t =0) 0.005 0.006 0.082 0.104<br />

(0.505) (0.375) (0.233) (0.118)<br />

(continued)<br />

1027<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/<br />

at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

1028<br />

Table A4<br />

Continued<br />

Increase in Sharpe ratio Increase in MPPM<br />

Dependent variable (9) (10) (11) (12) (13) (14) (15) (16)<br />

Share of tax-free assets (t =0) 0.063∗∗∗ 0.064∗∗∗ 0.074 0.156∗∗ (0.000) (0.000) (0.385) (0.036)<br />

Alpha (t = −44 to t =0) −3.405∗∗∗ −93.753∗∗∗ (0.000) (0.000)<br />

Constant 0.199∗∗∗ 0.138∗∗∗ 0.149∗∗∗ 0.133∗∗∗ 0.330∗∗∗ 0.433∗∗∗ 0.367∗∗∗ −0.034<br />

(0.000) (0.000) (0.000) (0.000) (0.000) (0.000) (0.003) (0.814)<br />

Observations 5,479 5,476 5,470 5,469 5,487 5,484 5,478 5,477<br />

R2 0.010 0.022 0.076 0.083 0.000 0.001 0.019 0.203<br />

Table A4 reports OLS estimates of the coefficients related to a decrease in HHI (models 1 to 4) and a decrease in idiosyncratic variance share (models 5 to 8), an increase in Sharpe ratio<br />

(models 9 to 12) and an increase in MPPM (models 13 to 16). HHI and portfolio returns are calculated based on the actual portfolios of investors. The focus of the table is on the variable<br />

Improvement of the alternative degree of following, which shows the improvement of the alternative degree of following from the time of the first recommendation to t = 11 and takes the<br />

value zero for non-advised clients. Dummy advice is equal to one if a client opts for financial advice. Additionally, the model controls for the following independent variables: a dummy<br />

variable that is equal to one if a client is male (Dummy male), the age of a client (Age), a dummy that is equal to one if a client falls into categories one to three of a microgeographic status<br />

rating by an external agency (Dummy low wealth), a dummy that is equal to one if a client falls into categories seven to nine of the microgeographic status (Dummy high wealth), the risky<br />

portfolio value of the customer (Log portfolio value), the number of years the client has been with the bank (Length of relationship), the number of trades per month (Trades per month), the<br />

average portfolio turnover per month (Portfolio turnover), the difference between the proportion of realized gains and losses (Disposition effect), the proportion of risky assets in the account<br />

(Risky share), the proportion of tax-free assets (Share of tax-free assets), and the weekly alpha of a particular customer before opting for financial advice (Alpha). Alpha and idiosyncratic<br />

variance share stem from the application of a Carhart (1997) four-factor model that is calibrated for Germany. P-values are in parentheses. R2 values and number of observations are<br />

reported. *** denotes significance at 1% or less, ** significance at 5% or less, and * significance at 10% or less. Heteroscedasticity robust standard errors are used. Different counts of<br />

observations are due to data availability of certain variables (see Table 5); results are robust to using the lowest common denominator.<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

Figure A1<br />

Initial alternative degree of following (buy side only). Counterparts of Figure 2<br />

Panel A: The cross-sectional distribution of the degree of following when the first recommendation was given.<br />

Panel B: The cross-sectional distribution of the average degree of following between time t = advice start and<br />

t = 11 months. Panel C: Increase in degree of following from the time of the first recommendation to t = 11.<br />

References<br />

Agarwal, S., J. <strong>Dr</strong>iscoll, X. Gabaix, and D. Laibson. 2009. The Age of Reason: Financial Decisions over the<br />

Life-cycle with Implications for Regulation. Brooking Papers on Economic Activity Fall:51–117.<br />

Agnew, J., P. Balduzzi, and A. Sundén. 2003. Portfolio Choice and Trading in a Large 401(k) Plan. American<br />

Economic Review 93:193–215.<br />

Ahearne, A. G., W. L. Griever, and F. E. Warnock. 2004. Information Costs and Home Bias: An Analysis of US<br />

Holdings of Foreign Securities. Journal of International Economics 62:331–36.<br />

Barber, B. M., Y.-T. Lee, Y.-J. Liu, and T. Odean. 2009. Just How Much Do Individual Investors Lose by<br />

Trading? Review of Financial Studies 22:609–32.<br />

Barber, B. M., R. Lehavy, M. McNichols, and B. Trueman. 2001. Can Investors <strong>Prof</strong>it from the Prophets?<br />

Security Analysts Recommendations and Stock Returns. Journal of <strong>Finance</strong> 56:531–63.<br />

Barber, B. M., and T. Odean. 2000. Trading Is Hazardous to Your Wealth: The Common Stock Investment<br />

Performance of Individual Investors. Journal of <strong>Finance</strong> 55:773–806.<br />

———. 2001. Boys Will Be Boys: Gender, Overconfidence, and Common Stock Investment. Quarterly Journal<br />

of Economics 116:261–92.<br />

Barras, L., O. Scaillet, and R. Wermers. 2010. False Discoveries in Mutual Fund Performance: Measuring Luck<br />

in Estimated Alphas. Journal of <strong>Finance</strong> 65:179–216.<br />

Benartzi, S., and R. Thaler. 2004. Save More Tomorrow: Using Behavioral Economics to Increase Employee<br />

Saving. Journal of Political Economy 112:164–87.<br />

1029<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

———. 2007. Heuristics and Biases in Retirement Savings Behavior. Journal of Economic Perspectives 21:<br />

81–104.<br />

Bergstresser, D., J. M. R. Chalmers, and P. Tufano. 2009. Assessing the Costs and Benefits of Brokers in the<br />

Mutual Fund Industry. Review of Financial Studies 22:4129–4156.<br />

Bertrand, M., and A. Morse. 2009. Information Disclosure, Cognitive Biases, and Payday Borrowing. Working<br />

Paper, Milton Friedman Institute for Research in Economics.<br />

Bhattacharya, U., C. Holden, and S. Jacobsen. Forthcoming. Penny Wise, Dollar Foolish: Threshold Trigger<br />

Effects in Security Trading. Management Science.<br />

Blume, M., and I. Friend. 1975. The Allocation of Wealth to Risky Assets—The Asset Structure of Individual<br />

Portfolios and Some Implications for Utility Functions. Journal of <strong>Finance</strong> 30:585–603.<br />

Bolton, P., X. Freixas, and J. Shapiro. 2007. Conflicts of Interest, Information Provision, and Competition in the<br />

Financial Services Industry. Journal of Financial Economics 85:297–330.<br />

Bonaccio, S., and R. S. Dalal. 2006. Advice Taking and Decision-making: An Integrative Literature Review, and<br />

Implications for the Organizational Sciences. Organizational Behavior and Human Decision Processes 101:<br />

127–51.<br />

———. 2010. What Types of Advice Do Decision-makers Prefer? Organizational Behavior and Human<br />

Decision Processes 112:11–23.<br />

Calvet, L. E., J. Y. Campbell, and P. Sodini. 2007. Down or Out: Assessing the Welfare Costs of Household<br />

Investment Mistakes. Journal of Political Economy 115:707–47.<br />

———. 2009a. Fight or Flight? Portfolio Rebalancing by Individual Investors. Quarterly Journal of Economics<br />

124:301–48.<br />

———. 2009b. Measuring the Financial Sophistication of Households. American Economic Review 99:393–98.<br />

Campbell, J. Y. 2006. Household <strong>Finance</strong>. Journal of <strong>Finance</strong> 61:1553–1604.<br />

Campbell, J. Y., H. Jackson, B. C. Madrian, and P. Tufano. 2011. Consumer Financial Protection. Journal of<br />

Economic Perspectives 25:91–114.<br />

Carhart, M. M. 1997. On Persistence in Mutual Fund Performance. Journal of <strong>Finance</strong> 52:57–82.<br />

Carlin, B. I. 2009. Strategic Price Complexity in Retail Financial Markets. Journal of Financial Economics<br />

91:278–87.<br />

Carlin, B. I., and S. Gervais. 2009. Legal Protection in Retail Financial Markets. Working Paper, University of<br />

California and Duke University.<br />

Carlin, B. I., and D. T. Robinson. 2009. Fear and Loathing in Las Vegas: Evidence from Blackjack Tables.<br />

Judgment and Decision Making 4:385–96.<br />

———. 2010. What Does Financial Literacy Training Teach Us? Working Paper, National Bureau of Economic<br />

Research.<br />

Carlin, B., and G. Manso. 2011. Obfuscation, Learning, and the Evolution of Investor Sophistication. Review of<br />

Financial Studies 24:754–85.<br />

CFA Institute. 2009. European Union Member Poll on Retail Investment Products. Summary Report.<br />

Chen, J., J. Kubik, and H. Hong. 2011. Outsourcing Mutual Fund Management: Firm Boundaries, Incentives,<br />

and Performance. Working Paper, University of California, Syracuse University, and Princeton University.<br />

Christelis, D., T. Jappelli, and M. Padula. 2010. Cognitive Abilities and Portfolio Choice. European Economic<br />

Review 54:18–38.<br />

Cooper, I., and E. Kaplanis. 1994. Home Bias in Equity Portfolios, Inflation Hedging, and International Capital<br />

Market Equilibrium. Review of Financial Studies 7:45–60.<br />

1030<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Is Unbiased Advice Sufficient?<br />

DABbank. 2004. Faszination Wertpapier: Fakten und Hintergründe zum Anlegerverhalten in Deutschland. DAB-<br />

Bank: Munich.<br />

Deaves, R., E. Lüders, and G. Y. Luo. 2003. An Experimental t-test of the Impact of Overconfidence and Gender<br />

on Trading Activity. Review of <strong>Finance</strong> 13:555–75.<br />

DeMiguel, V., L. Garlappi, and R. Uppal. 2009. Optimal versus Naive Diversification: How Inefficient Is the<br />

1/N Portfolio Strategy? Review of Financial Studies 22:1915–1953.<br />

Deuskar, P., J. M. Pollet, Z. J. Wang, and L. Zheng. 2011. The Good or the Bad? Which Mutual Fund Managers<br />

Join Hedge Funds? Review of Financial Studies 24:3008–24.<br />

Deutsche Bundesbank. 2010. Wertpapierdepots März 2010. Frankfurt am Main.<br />

Deutsches Aktieninstitut. 2009. DAI Factbook 2009: Statistiken, Analysen und Graphiken zu Aktionären,<br />

Aktiengesellschaften und Börsen. Frankfurt am Main.<br />

DiMatteo, M. R. 2004. Variations in Patient’s Adherence to Medical Recommendations: A Quantitative Review<br />

of 50 Years of Research. Medical Care 42:200–209.<br />

Dimson, E., P. Marsh, and M. Staunton. 2007. The Worldwide Equity Premium: A Smaller Puzzle. In R. Mehra<br />

(ed.), Handbook of Investments: Equity Risk Premium, pp. 467–514. Amsterdam: Elsevier.<br />

Dorn, D., G. Huberman, and P. Sengmueller. 2008. Correlated Trading and Returns. Journal of <strong>Finance</strong> 63:<br />

885–920.<br />

Edelen, R. M., R. Evans, and G. B. Kadlec. 2008. What Do Soft-dollars Buy? Performance, Expense Shifting,<br />

Agency Costs. Working Paper, AFA San Francisco Meetings.<br />

Fama, E. F., and K. R. French. 1993. Common Risk Factors in the Returns on Stocks and Bonds. Journal of<br />

Financial Economics 33:3–56.<br />

——. 2010. Luck versus Skill in the Cross-section of Mutual Fund Returns. Journal of <strong>Finance</strong> 65:1915–1947.<br />

Frazzini, A. 2006. The Disposition Effect and Underreaction to News. Journal of <strong>Finance</strong> 61:2017–2046.<br />

French, K., and J. Poterba. 1991. Investor Diversification and International Equity Markets. American Economic<br />

Review 81:222–26.<br />

Gino, F. 2008. Do We Listen to Advice Just Because We Paid for It? The Impact of Advice Cost on Its Use.<br />

Organizational Behavior and Human Decision Processes 107:234–45.<br />

Goetzmann, W., J. <strong>Ing</strong>ersoll, M. Spiegel, and I. Welch. 2007. Portfolio Performance Manipulation and<br />

Manipulation-proof Performance Measures. Review of Financial Studies 20:1503–1546.<br />

Goetzmann, W., and A. Kumar. 2008. Equity Portfolio Diversification. Review of <strong>Finance</strong> 12:433–63.<br />

Guiso, L., M. Haliassos, and T. Jappelli. 2002. Household Portfolios. Cambridge, MA: MIT Press.<br />

Hackethal, A., M. Haliassos, and T. Jappelli. 2012. Financial Advisors: A Case of Babysitters? Journal of<br />

Banking & <strong>Finance</strong> 36:509–24.<br />

Heywood, J. 1546. A Dialogue Conteinyng the Nomber in Effect of all the Prouerbes in the Englishe Tongue.<br />

Classic Books.<br />

Holden, C. 2009. Excel Modeling and Estimation, 3rd ed. Upper Saddle River, NJ: Pearson Prentice Hall.<br />

Hong, H., J. D. Kubik, and J. C. Stein. 2005. Thy Neighbor’s Portfolio: Word-of-mouth Effects in the Holdings<br />

and Trades of Money Managers. Journal of <strong>Finance</strong> 60:2801–2824.<br />

Huberman, G. 2001. Familiarity Breeds Investment. Review of Financial Studies 14:659–80.<br />

Inderst, R., and M. Ottaviani. 2009. How (Not) to Pay for Advice: A Framework for Consumer Protection.<br />

Working Paper, Goethe University Frankfurt and Northwestern University.<br />

Investment Company Institute. 2007. Why Do Mutual Fund Investors Use <strong>Prof</strong>essional Financial Advisors?<br />

Research Fundamentals 16:1–8.<br />

1031<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


The Review of Financial Studies / v 25 n 4 2012<br />

Ivkovic, Z., C. Sialm, and S. Weisbenner. 2008. Portfolio Concentration and the Performance of Individual<br />

Investors. Journal of Financial and Quantitative Analysis 43:613–55.<br />

Jorion, P. 1986. Bayes-Stein Estimation for Portfolio Analysis. Journal of Financial and Quantitative Analysis<br />

21:279–92.<br />

Kahneman, D. 2003. Maps of Bounded Rationality: Psychology for Behavioral Economics. American Economic<br />

Review 93:1449–1475.<br />

Kelly, M. 1995. All Their Eggs in One Basket: Portfolio Diversification of US Households. Journal of Economic<br />

Behavior and Organization 27:87–96.<br />

Laibson, D. 1997. Golden Eggs and Hyperbolic Discounting. Quarterly Journal of Economics 112:443–77.<br />

Lewis, K. K. 1999. Trying to Explain Home Bias in Equities and Consumption. Journal of Economic Literature<br />

37:571–608.<br />

Madrian, B. C., and D. F. Shea. 2001. The Power of Suggestion: Inertia in 401(k) Participation and Savings<br />

Behavior. Quarterly Journal of Economics 116:1149–1187.<br />

Malkiel, B. 1995. Returns from Investing in Equity Mutual Funds 1971 to 1991. Journal of <strong>Finance</strong> 50:549–72.<br />

Markowitz, H. M. 1952. Portfolio Selection. Journal of <strong>Finance</strong> 7:77–91.<br />

Mehra, R., and E. C. Prescott. 1985. The Equity Premium: A Puzzle. Journal of Monetary Economics 15:145–61.<br />

Metrick, A. 1999. Performance Evaluation with Transactions Data: The Stock Selection of Investment<br />

Newsletters. Journal of <strong>Finance</strong> 54:1743–1774.<br />

Michaud, R. 1998. Efficient Asset Management: A Practical Guide to Stock Portfolio Optimization and Asset<br />

Allocation. New York: Oxford University Press.<br />

Mullainathan, S., M. Nöth, and A. Schoar. 2009. The Market for Financial Advice: An Audit Study. Working<br />

Paper, AFA Denver Meeting.<br />

NASD Investor Education Foundation. 2006. Investor Fraud Study: Final Report.<br />

O’Donoghue, T., and M. Rabin. 1999. Doing It Now or Later. American Economic Review 89:103–24.<br />

Odean, T. 1998. Are Investors Reluctant to Realize Their Losses? Journal of <strong>Finance</strong> 53:1775–1798.<br />

———. 1999. Do Investors Trade Too Much? American Economic Review 89:1278–1298.<br />

Poterba, J. M., and A. A. Samwick. 2003. Taxation and Household Portfolio Composition: US Evidence from<br />

the 1980s and 1990s. Journal of Public Economics 87:5–38.<br />

Puri, M., and D. T. Robinson. 2007. Optimism and Economic Choice. Journal of Financial Economics 86:71–99.<br />

Samuelson, W., and R. J. Zeckhauser. 1988. Status Quo Bias in Decision-making. Journal of Risk and<br />

Uncertainty 1:7–59.<br />

Sharpe, W. F. 1966. Mutual Fund Performance. Journal of Business 39:119–38.<br />

Shefrin, H., and M. Statman. 1985. The Disposition to Sell Winners Too Early and Ride Losers Too Long:<br />

Theory and Evidence. Journal of <strong>Finance</strong> 40:777–90.<br />

Statman, M. 2010. What Investors Really Want. New York: McGraw-Hill.<br />

Stoughton, N. M., Y. Wu, and J. Zechner. 2011. Intermediated Investment Management. Journal of <strong>Finance</strong><br />

66:947–80.<br />

Womack, K. L. 1996. Do Brokerage Analysts’ Recommendations Have Investment Value? Journal of <strong>Finance</strong><br />

51:137–67.<br />

Yaniv, I. 2004. Receiving Other People’s Advice: Influence and Benefit. Organizational Behavior and Human<br />

Decision Processes 93:1–13.<br />

Zhu, N. 2002. The Local Bias of Individual Investors. Working Paper, Yale.<br />

1032<br />

Downloaded from<br />

http://rfs.oxfordjournals.org/ at Universitaetsbibliothek Johann Christian Senckenberg on July 3, 2012


Financial advisors: A case of babysitters?<br />

Andreas Hackethal a , Michael Haliassos a,c,d , Tullio Jappelli b,c,e,⇑<br />

a House of <strong>Finance</strong>, Goethe University Frankfurt, Grueneburgplatz 1, D-60323 Frankfurt am Main, Germany<br />

b Department of Economics, University of Naples Federico II, Via Cinzia 45, 80126 Naples, Italy<br />

c CEPR, Centre for Economic Policy Research, London, United Kingdom<br />

d CFS, Center for Financial Studies, Frankfurt, Germany<br />

e CSEF, Centre for Studies in Economics and <strong>Finance</strong>, Naples, Italy<br />

article info<br />

Article history:<br />

Received 1 March 2011<br />

Accepted 31 August 2011<br />

Available online 8 September 2011<br />

JEL classification:<br />

G1<br />

E2<br />

D8<br />

Keywords:<br />

Financial advice<br />

Portfolio choice<br />

Household finance<br />

1. Introduction<br />

abstract<br />

In recent years households have increased their exposure to<br />

financial risk taking, partly in response to the demographic transition<br />

and increased responsibility for retirement financing. Recent<br />

research points to differential financial literacy and sophistication<br />

across households, creating the potential for important distributional<br />

consequences of these developments (Campbell, 2006; Lusardi<br />

and Mitchell, 2007).<br />

In principle, financial advisors could ameliorate consequences<br />

of differential ability to handle finances by improving returns<br />

and ensuring greater risk diversification among less sophisticated<br />

households. Indeed, delegation of portfolio decisions to advisors<br />

opens up economies of scale in portfolio management and information<br />

acquisition, because advisors can spread information<br />

acquisition costs among many investors. Such economies of scale,<br />

as well as possibly superior financial practices of advisors, create<br />

the potential for individual investors to improve portfolio performance<br />

by delegating financial decisions. But delegation entails<br />

⇑ Corresponding author at: Department of Economics, University of Naples<br />

Federico II, Via Cinzia 45, 80126 Naples, Italy. Tel.: +39 081 675042; fax: +39 081<br />

675014.<br />

E-mail addresses: hackethal@gbs.uni-frankfurt.de (A. Hackethal), haliassos@wi<br />

wi.uni-frankfurt.de (M. Haliassos), tullio.jappelli@unina.it (T. Jappelli).<br />

0378-4266/$ - see front matter Ó 2011 Elsevier B.V. All rights reserved.<br />

doi:10.1016/j.jbankfin.2011.08.008<br />

Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

Contents lists available at SciVerse ScienceDirect<br />

Journal of Banking & <strong>Finance</strong><br />

journal homepage: www.elsevier.com/locate/jbf<br />

We use two data sets, one from a large brokerage and another from a major bank, to ask: (i) whether<br />

financial advisors are more likely to be matched with poorer, uninformed investors or with richer and<br />

experienced investors; (ii) how advised accounts actually perform relative to self-managed accounts;<br />

(iii) whether the contribution of independent and bank advisors is similar. We find that advised accounts<br />

offer on average lower net returns and inferior risk-return tradeoffs (Sharpe ratios). Trading costs contribute<br />

to outcomes, as advised accounts feature higher turnover, consistent with commissions being the<br />

main source of advisor income. Results are robust to controlling for investor and local area characteristics.<br />

The results apply with stronger force to bank advisors than to independent financial advisors, consistent<br />

with greater limitations on bank advisory services.<br />

Ó 2011 Elsevier B.V. All rights reserved.<br />

costs in terms of commissions and fees, and might give rise to<br />

agency problems between advisors and firms and between advisors<br />

and poorly informed customers, as shown by Inderst and Ottaviani<br />

(2009): on the one hand they need to sell financial products<br />

and on the other they need to advise customers on what is best for<br />

them to do. 1 The notion that financial advisors tend to be used by<br />

less informed or unsophisticated investors who could be easily misled<br />

by them, underlies much of the existing literature on financial literacy,<br />

the possible role of financial advice and the case for regulation<br />

of financial advisors.<br />

In this paper, we examine three questions. First, we ask whether<br />

financial advisors tend to be matched with poorer, uninformed<br />

investors or rather with richer, experienced investors with higher<br />

opportunity cost of time. Second, how brokerage accounts run by<br />

individuals without financial advisors actually perform compared<br />

to accounts run by (or in consultation with) financial advisors.<br />

Third, whether the estimated contribution of financial advisors is<br />

persistent across different advisory models such as independent<br />

financial advisors (IFA) and bank financial advisors (BFA). Direct<br />

performance comparisons are made possible by two unique<br />

administrative data sets: one from a large German brokerage firm<br />

that allows its clients choice of whether to run their accounts<br />

1 Stulz and Mehran (2007) review the existing empirical literature on the nature<br />

and implications of various conflicts mainly focusing on analysts.


510 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

themselves or with the guidance of an independent financial advisor;<br />

and another from a large German commercial bank that offers<br />

(optional) advice to its customers with investment accounts. The<br />

answers we obtain provide a novel perspective on the role of financial<br />

advice in individual portfolio performance.<br />

Our first data set, from the online brokerage, tracks accounts of<br />

32,751 randomly selected individual customers. Our second data<br />

set contains full data of 4447 clients of a German branch-based<br />

commercial bank. Both data sets cover 34 months, from January<br />

2003 to October 2005. In many respects, discussed in this paper,<br />

descriptive statistics for both samples paint a very similar picture<br />

of the role of financial advisors as econometric analysis that controls<br />

for investor and region characteristics.<br />

We find that involvement of financial advisors lowers portfolio<br />

returns net of direct cost, worsens risk-return profiles, as measured<br />

by the Sharpe ratio; and increases account turnover and investment<br />

in mutual funds, consistent with incentives built into the<br />

commission structure of both types of financial advisors. If anything,<br />

negative advisory effects on portfolio performance are even<br />

stronger for BFAs than for IFAs. This is consistent with greater limitations<br />

faced by BFAs in the range of products they offer and in the<br />

way they can confer financial advice to clients.<br />

Regression analysis of who delegates portfolio decisions presents<br />

a further twist. It suggests that advisors are matched with richer,<br />

older, more experienced, self-employed, female investors<br />

rather than with poorer, younger, inexperienced and male ones.<br />

In this respect, advisors are similar to babysitters: babysitters are<br />

matched with well-to-do parents, they perform a service that parents<br />

themselves could do better, they charge for it, but observed<br />

child achievement is not boosted by babysitters but by positive<br />

characteristics of the family. No issues of regulating babysitters<br />

emerge, however, because the nature of the activity and the contribution<br />

is known to all parties involved.<br />

The paper is organized as follows. In Section 2 we discuss the<br />

role of financial advice in overcoming investors’ informational constraints<br />

and their incentives in handling financial portfolios in view<br />

of relevant existing literature. Section 3 describes the brokerage<br />

and the commercial bank data sets, the measures that we use to<br />

characterize portfolio performance, and the estimation procedure.<br />

Section 4 compares records of account performance with and without<br />

involvement of financial advisors. Section 5 studies econometrically<br />

the role of investors’ characteristics in determining which<br />

investors are matched with financial advisors. Section 6 reports<br />

regression estimates of the effects of independent financial advisors<br />

on account performance, return variance, Sharpe ratios, trading<br />

frequency, turnover, and diversification. Section 7 presents similar<br />

results from the second sample on bank financial advisors. Section<br />

8 concludes.<br />

2. The role of financial advice<br />

There is a limited but budding theoretical literature on the possible<br />

role of financial advisors. Current theoretical work but also<br />

policy debate on financial regulation seem to be based on the idea<br />

that financial advisors know what is good for individual customers<br />

but have an incentive to misrepresent this and to take advantage of<br />

their customers, who are typically uninformed and cannot figure<br />

out the poor quality of advice. Regulation is then needed to make<br />

sure that this conflict of interests is dealt with.<br />

In a recent pioneering paper, Inderst and Ottaviani (2009) analyze<br />

‘misselling’, i.e. the practice of misdirecting clients into buying<br />

a financial product that is not suitable for them. Their model<br />

emphasizes the internal agency problem between the firm and<br />

its sales agents. The agency problem is complicated by the fact that<br />

sales agents perform the dual task of prospecting for customers<br />

and of providing adequate advice to them on whether to buy a particular<br />

product. As a consequence, higher sales incentives will increase<br />

the likelihood that sales agents sell unsuitable products to<br />

customers. If this occurs, there is a probability with which the firm<br />

receives a complaint and has to pay a fine. To avoid misselling the<br />

firm can set internal suitability standards for advising customers<br />

and exert costly monitoring to verify compliance with these standards.<br />

The standards implemented by the firm in equilibrium are<br />

increasing in the fine (or equivalently in the reputation damage),<br />

the transparency of the incentive scheme, and in the effectiveness<br />

of monitoring, but they are decreasing in the sales incentives and<br />

the private cost for the agent to investigate the match between<br />

product and customer.<br />

There are two relevant implications for our study. First, sales<br />

incentives can lead financial advisors to systematically recommend<br />

unsuitable products to their clients that entail suboptimal outcomes<br />

on the client side. Second, due to agency costs from multitasking<br />

and monitoring, a firm employing sales agents (such as<br />

BFAs) would be expected to choose lower standards than an entrepreneur<br />

(IFA). Our findings below are quite consistent with these<br />

predictions and provide two further insights: (i) advisors may affect<br />

portfolio outcomes not only by recommending unsuitable<br />

products but also by encouraging excessive trading; and (ii) the notion<br />

that advisors have an edge over their clients need not refer solely<br />

to unsophisticated clients, but also to experienced but<br />

inattentive ones who fail to monitor advisors and the outcome of<br />

their activities effectively.<br />

The empirical literature on financial advice has so far mostly focused<br />

on whether professional analysts and advisors have an informational<br />

advantage to contribute to individual investors when it<br />

comes to predicting stock price movements. Ever since Cowles<br />

(1933), there have been questions regarding the ability of stock<br />

market forecasters and analysts to predict and reveal movements<br />

in the stock market. 2<br />

For example, Womack (1996) examines stock price movements<br />

following ‘buy’ or ‘sell’ recommendations by fourteen major US<br />

brokerage firms. He documents significant price and volume reactions<br />

in the direction of the recommendation, especially for new<br />

‘sell’ recommendations. He concludes that there is value to these<br />

recommendations viewed as returns to information search costs.<br />

However, new ‘buy’ recommendations occur seven times more often<br />

than ‘sell’ recommendations, suggesting that brokers are reluctant<br />

to issue sell recommendations, both in order to avoid harming<br />

potential investment banking relationships and to maintain future<br />

information flows from managers.<br />

Metrick (1999) analyzes a database of recommendations of 153<br />

investment newsletters and finds no evidence that newsletters<br />

have superior stock-selection skill. Average abnormal returns are<br />

close to zero; and even the performance of the best newsletters<br />

seems to be driven more by luck then by skill. In related work,<br />

Anderson and Martinez (2008) examine abnormal returns around<br />

stock recommendations by Swedish brokers. A sizeable share of<br />

abnormal profits results from transactions before the recorded recommendation<br />

date, suggesting that tipping of customers may be<br />

taking place. However, given the small size of these abnormal profits<br />

(only 0.04% in yearly performance of total Swedish equity fund<br />

assets under management), the authors wonder whether clients<br />

are fully compensated for the costs of commissions charged by<br />

brokers.<br />

Barber et al. (2001) explicitly take into account trading costs<br />

from following analyst recommendations. They analyze abnormal<br />

gross and net returns that would result from purchasing (selling<br />

2 Early studies include Barber and Loeffler (1993) on The Wall Street Journal’s<br />

Dartboard column and Desai and Jain (1995) on ‘‘Superstar’’ money managers in<br />

Barron’s.


short) stocks with the most (least) favorable consensus recommendations,<br />

in conjunction with daily portfolio rebalancing and a<br />

timely response to recommendation changes. Although they find<br />

that such strategies would yield annual abnormal gross returns<br />

greater than 4%, they also show that abnormal net returns are<br />

not statistically significant. Bergstresser et al. (2009) compare performance<br />

of mutual fund ‘classes’ distinguished by their distribution<br />

channel: directly sold to investors versus sold through<br />

brokers, with correspondingly different fee structures. They find<br />

that funds sold through brokers offer inferior returns, even before<br />

the distribution fee, no superior aggregate market timing ability,<br />

and exhibit the same return-chasing behavior as observed among<br />

direct-channel funds. Finally, more sales are directed to funds with<br />

larger distribution fees.<br />

Our reading of the literature on informational contributions of<br />

analysts or brokers to direct stockholding is that these may be<br />

present but unlikely to be exploitable by individuals given the<br />

trading costs they entail. Therefore, in a world in which financial<br />

advisors solely provided security selection advice we would expect<br />

the effect of financial advice on abnormal portfolio returns to be<br />

around zero on average after transaction costs.<br />

However, some researchers take a different angle and point out<br />

that, even if professional advisors do not have superior information<br />

that is exploitable for the normal trading within an individual<br />

account, they may be less likely to exhibit behavioral biases that<br />

hurt account performance. They could thus help either by running<br />

the account themselves or by encouraging investors to behave<br />

appropriately.<br />

A behavioral bias that has received considerable attention is the<br />

‘disposition effect’, i.e. the tendency of some individuals to sell<br />

winners and keep losers when it comes to direct stockholding<br />

(Odean, 1998). Shapira and Venezia (2001) found that the disposition<br />

effect is significantly less pronounced among professional<br />

than among self-directed investors. Well trained advisors could<br />

therefore aid their clients in reducing the disposition effect, potentially<br />

enhancing risk adjusted portfolio returns.<br />

Advisors might also be simply able to moderate trading activity<br />

(Campbell and Viceira, 2003). Barber and Odean (2000) show that<br />

some investors trade excessively in brokerage accounts, suffering<br />

transactions costs that result in significantly lower returns. Such<br />

behavior is often attributed to overconfidence, especially pronounced<br />

among male investors (Odean, 1999; Barber and Odean,<br />

2001). Shu et al. (2004) analyze the returns on common stock<br />

investments by 52,649 accounts at a brokerage house in Taiwan<br />

for 45 months ending in September 2001. They find a U-shaped<br />

turnover and performance relation rather than the monotonic<br />

one predicted by overconfidence: the most frequent traders in<br />

the top turnover quintile perform better than investors in the middle<br />

three quintiles. Other behavioral biases have been found to<br />

influence some individual investors. For instance, Venezia et al.<br />

(2011) focus on herding, and also document that professional<br />

investors herd less than amateurs, while Grinblatt and Keloharju<br />

(2001) suggest that investors trade on the basis of past returns, reference<br />

prices, or the size of holding period gain or loss.<br />

While the list of potential behavioral biases can grow longer, an<br />

important question – consistent with our approach in this paper –<br />

remains as to whether individuals who exhibit such biases are<br />

likely to make use of professional investors. For example, Guiso<br />

and Jappelli (2006) argued that overconfidence (i.e. the disposition<br />

of investors to overstate the value of their private information) reduces<br />

their propensity to seek advice. Indeed, the Barber and<br />

Odean data come from a discount broker that does not offer advice.<br />

Even if overconfident traders approach financial advisors, one<br />

might wonder whether financial advisors who earn sales commissions<br />

would actually discourage them from executing too many<br />

trades without some incentive scheme.<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 511<br />

On the other hand, financial advisors may help correct behavioral<br />

biases or investment mistakes when such correction is<br />

aligned with their interests. A case in point is diversification. A<br />

number of empirical studies find that many individual investors<br />

hold undiversified portfolios (see e.g. Campbell, 2006; Goetzmann<br />

and Kumar, 2008). Financial advisors who earn commissions for<br />

selling mutual funds have an incentive to promote such sales and<br />

through them diversification of their client’s accounts. Shapira<br />

and Venezia (2001) find that the number of different stocks and<br />

the number of transactions per year is about three times as high<br />

for accounts managed by professionals than for self-managed accounts.<br />

This finding indicates that financial advisors might promote<br />

diversification through single stocks if they participate in<br />

brokerage commissions. 3<br />

Taken together, literature suggests an ambiguous effect of<br />

financial advice on net returns and risk profiles of client portfolios.<br />

Although it seems to be rather unlikely that advisors enhance portfolio<br />

performance through informational contributions they might<br />

in fact improve the risk-return profile by ironing out behavioral<br />

biases of their clients. Of course, such positive effects must exceed<br />

the cost of advice in order to yield an overall positive effect.<br />

Our paper takes a direct approach to the issue of the role and<br />

contribution of financial advisors. Recognizing both the potential<br />

informational advantage and the potential contribution of professional<br />

investors to controlling behavioral biases and correcting<br />

investment mistakes, it compares directly portfolio returns (net<br />

of transactions costs) and portfolio risk levels that investors actually<br />

accomplish on their own versus what they accomplish with<br />

the guidance of a financial advisor. It does so with reference to<br />

portfolios actually chosen and adjusted by investors, which include<br />

directly held stocks, bonds, and mutual funds; and it accounts for a<br />

number of investors’ and region characteristics observable in our<br />

data and for how they influence the tendency to use a financial<br />

advisor. Moreover, we are able to measure the effects of advice<br />

across two distinct advisory models (IFAs and BFAs).<br />

3. Data, measurement, and estimation<br />

3.1. Data on independent financial advisors<br />

The first data set we use is administrative information from a<br />

large German brokerage firm. It covers the investments of 32,751<br />

randomly selected customers. They all had an active account with<br />

the brokerage firm over the sample period from January 2003 to<br />

October 2005. If customers opened multiple accounts we consolidated<br />

them into one single account.<br />

For each sampled customer we have information on date of<br />

birth, gender, marital status, profession (including status as employed<br />

or self-employed), zip-code of place of residence, nationality,<br />

and self-reported security-trading experience in years. 4 All<br />

information was collected by the brokerage firm on the date of account<br />

opening and updated according to new information that the<br />

firm has obtained from the customer in the interim.<br />

On average (not excluding account owners aged under 18),<br />

sample customers held 38.6% of their account volume in the form<br />

of equity mutual funds, 47.4% in the form of single stocks (28%<br />

3 Moreover, Shapira and Venezia (2001) find that the round trip performance of<br />

professionally managed accounts is slightly (and for some specifications also<br />

statistically) higher than that of independently run accounts. The discrepancy to<br />

our own results is likely due to the fact that they focus on round trip returns for stock<br />

investments in one particular year, whereas we measure total portfolio returns for a<br />

longer time period, and in addition control for individual investor characteristics.<br />

4 Self-reported trading experience is reported on a scale with intervals equal to 5<br />

years. We construct a variable that has the interval midpoints as values and then add<br />

the number of years that elapsed since account opening to approximate total trading<br />

experience at the beginning of our observation period.


512 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

thereof in German stocks), 2.4% in the form of bond mutual funds,<br />

3.8% in the form of single bonds and the remainder in the form of<br />

structured investment certificates, warrants, and other securities.<br />

Our administrative data set includes a variable that indicates<br />

whether a given brokerage customer is also a client of an IFA<br />

who registered with the brokerage firm. We know from the brokerage<br />

firm that, typically, advised customers were brought to the<br />

brokerage by IFAs. About 90% of IFAs registered with the brokerage<br />

are former employees of commercial banks advising customers on<br />

investment accounts. They decided to leave the bank and become<br />

independent, thereby offering lower costs than banks and greater<br />

choice of financial products. Thus, they were able to persuade<br />

many of their former customers at the bank to transfer investment<br />

accounts to the brokerage firm. The remaining 10% of IFAs in our<br />

sample are not former bank employees but they instead joined a<br />

larger team of IFAs directly and built up their own customer base,<br />

again drawing mostly from former bank customers.<br />

At the time of account opening, IFAs had typically obtained a<br />

client mandate to place orders on behalf of the client. We do not<br />

have information on which clients fully delegate trading decisions<br />

to their IFAs and which only consult their IFAs for guidance and<br />

then place trades themselves. The brokerage firm offers several<br />

compensation schemes to IFAs. Only for a negligible fraction of<br />

IFAs are revenues dependent solely on assets under management.<br />

More than 90% of IFAs generate at least a portion of revenues from<br />

trades, such as sales commissions. In the case of mutual funds the<br />

commission is a function of the upfront load the brokerage firm<br />

earns from the fund producer.<br />

Of the customers in our sample, 12.8% consult IFAs registered<br />

with the brokerage firm. More than half of these customers are<br />

IFAs’ former banking clients, with the remaining half (typically also<br />

former bank customers) having been acquired over the years, most<br />

importantly through existing customers’ referrals. We cannot rule<br />

out that customers coded as not using an IFA obtain professional<br />

advice from outside advisors. This is, however, rather unlikely because<br />

such outside advisors do not participate in the fees and commissions<br />

paid by the client to the brokerage firm and must<br />

therefore charge their services on top of the full brokerage fees<br />

and commissions.<br />

Table 1 shows descriptive statistics of the total brokerage sample<br />

and of the two sub-samples distinguished by whether sample<br />

customers were advised by IFAs or not, after dropping accounts<br />

that report age of account owner below 18. 5 As shown in the table,<br />

77.8% of account owners were male, and 47.9% married. Overall,<br />

86.1% were employed (excluding public servants) and 13.2% were<br />

self-employed, with the remaining 0.7% being public servants, retirees,<br />

housewives or students. Average trading experience as of January<br />

2003 was 9.34 years. Among IFA-assisted customers, men are<br />

underrepresented relative to their share in the overall account owner<br />

pool, and so are married owners (as indicated by the corresponding<br />

t-statistics and p-values in the last two columns of Table 1). Older<br />

owners (above 50) are overrepresented, and advised customers have<br />

on average more years of experience and larger initial size of<br />

accounts.<br />

Table 1 also reports performance figures for the brokerage accounts.<br />

As indicated by the t-statistics and associated p-values,<br />

raw returns, abnormal returns (see Section 3.3 for definitions),<br />

5 These are typically accounts run by parents on behalf of their children.<br />

Specifically, 796 investors in our original sample were younger than 18 on September<br />

5, 2006, and the youngest investor in that sample was just under 6 years old. Tax<br />

advantages for parents arise because during the observation period there was a per<br />

person threshold level of interest or dividend income above which capital income tax<br />

needed to be paid. We have also run the regressions including investors under 18, but<br />

our results were hardly affected in terms of sign, significance, and even size of<br />

estimates, except for small changes in the estimates for age categories.<br />

Table 1<br />

Descriptive statistics for the brokerage sample (IFAs).<br />

Selfmanaged<br />

accounts<br />

sample<br />

mean<br />

Accounts<br />

run by<br />

IFA<br />

sample<br />

mean<br />

T-test for<br />

difference<br />

in means<br />

All<br />

accounts<br />

sample<br />

mean<br />

All<br />

accounts<br />

standard<br />

deviation<br />

Dependent variables<br />

Log monthly<br />

returns<br />

0.0101 0.0063 24.70 *<br />

0.0097 0.0089<br />

Jensen’s alpha 0.0098 0.0061 23.72 *<br />

0.0093 0.0091<br />

Alpha – four<br />

factor<br />

model<br />

0.0093 0.0055 22.31 *<br />

0.0088 0.0100<br />

Variance of<br />

monthly<br />

returns<br />

0.0032 0.0019 28.39 *<br />

0.0031 0.0027<br />

Sharpe ratio 0.2229 0.1916 11.73 *<br />

0.2189 0.1585<br />

No. of trades/<br />

account<br />

volume in<br />

‘000<br />

0.0861 0.0884 0.87 0.0864 0.2609<br />

Monthly<br />

turnover<br />

rate<br />

0.0405 0.0895 33.38 *<br />

0.0468 0.0865<br />

Share of<br />

directly<br />

held stocks<br />

0.5777 0.2000 58.70 *<br />

0.5295 0.3838<br />

Control variables<br />

Male 0.7925 0.6739 15.37 *<br />

0.7774 0.4160<br />

Married 0.4812 0.4636 0.92 0.4790 0.4996<br />

Employed<br />

(excluding<br />

public<br />

servants)<br />

0.8655 0.8334 4.93 *<br />

0.8614 0.3455<br />

Selfemployed<br />

0.1280 0.1577 5.10 *<br />

0.1318 0.3383<br />

Experience 9.3415 11.1535 16.27 *<br />

9.5684 6.2182<br />

18 6 Age 6 30 0.0101 0.0415 3.43 *<br />

0.0462 0.2100<br />

30 < Age 6 40 0.0098 0.1180 18.35 *<br />

0.2409 0.4276<br />

40 < Age 6 50 0.0093 0.2680 8.66 *<br />

0.3346 0.4719<br />

50 < Age 6 60 0.0530 0.2287 5.00 *<br />

0.1995 0.3997<br />

Age > 60 0.0708 0.3437 28.11 *<br />

0.1787 0.3831<br />

Log account<br />

volume in<br />

2003<br />

9.1588 10.2823 42.98 *<br />

9.3023 1.4917<br />

Observations 25,173 3686 28,321 28,321<br />

The t-test refers to a test of the null hypothesis that the mean of the sample with<br />

self-managed accounts equals the mean of the sample with accounts run by an<br />

independent financial advisor (IFA).<br />

* The two means are statistically different from each other at the 1% level.<br />

raw return variance and Sharpe ratios are on average significantly<br />

lower for IFA-assisted than for self-directed customers.<br />

All reported return figures are monthly and net of any transactions<br />

costs and provisions charged by the brokerage on its own account<br />

or on behalf of the IFA. 6 Transaction costs and provisions are<br />

divided between the brokerage and IFA, with the bank typically<br />

earning roughly 30 basis points for transaction fees, account maintenance,<br />

and front loads, leaving about 170 basis points for the IFA.<br />

There is a minority of advisors who follow a different business model:<br />

instead of earning front loads, they forward those to their clients<br />

and earn an extra fee as a percentage of account volume. As this<br />

6 Although we only observe net returns in our data and therefore cannot directly<br />

measure transaction cost, we know from the data provider that the brokerage and the<br />

IFA combined earn typically 100–200 basis points on clients with account volume<br />

greater than 50,000 Euros. For smaller accounts, this number is typically in the<br />

neighborhood of 200 basis points, although it can be as high as 300–500 basis points,<br />

due to front loads (principally observable in the data set) and kick-backs (not<br />

observable) from mutual funds.


extra fee is not run through the bank, it is not observed by us and it is<br />

not taken into account in computing returns and other measures of<br />

performance net of costs. Since we obtain negative effects of IFAs on<br />

account performance in econometric estimations below, the resulting<br />

understatement of costs in these cases, if anything, strengthens<br />

our findings on the role of IFAs.<br />

The monthly position statements list for each item the type of<br />

security (e.g. stocks, bonds, mutual funds, etc.), the number of<br />

securities, and the market value per security at month end. At<br />

the start of the sample period (January 2003), average annual account<br />

volume was 10,963 Euro. We compute monthly turnover<br />

by dividing the combined transaction value of all purchase transactions<br />

for a given month by the average of beginning-of-month and<br />

end-of-month account volume. Average monthly turnover is 4.7%<br />

in our sample, but about double of this for advised customers.<br />

3.2. Data on bank financial advisors<br />

In order to compare our findings across different advisory models,<br />

we also consider a second data set of investment accounts, this<br />

time from a large German commercial bank that offers optional advice<br />

to its customers through its bank employees assigned to this<br />

task. Unlike the online brokerage that likely attracts a selected<br />

sample of the German population interested in trading, the bank<br />

has a wide network of branches that reach a broad cross section<br />

of the German population. This data set consists of 10,434 randomly<br />

selected customers observed over a 34-month period, from<br />

Table 2<br />

Descriptive statistics for the bank sample (BFAs).<br />

Selfmanaged<br />

accounts<br />

sample<br />

mean<br />

Accounts<br />

run by<br />

BFAs<br />

sample<br />

mean<br />

T-test for<br />

difference<br />

in means<br />

All<br />

accounts<br />

sample<br />

mean<br />

All<br />

accounts<br />

standard<br />

deviation<br />

Dependent variables<br />

Log monthly<br />

returns<br />

0.0076 0.0040 11.59 0.0054 0.0101<br />

Variance of<br />

monthly<br />

returns<br />

0.0045 0.0046 0.43 0.0046 0.0130<br />

Sharpe ratio 0.4252 0.2662 12.23 0.3253 0.4020<br />

Monthly<br />

turnover<br />

rate<br />

0.0680 0.0520 8.34 0.0579 0.1103<br />

Share of<br />

directly<br />

held stocks<br />

0.2975 0.1188 20.07 0.1853 0.3009<br />

Control variables<br />

Male 0.5102 0.4350 4.86 0.4630 0.4986<br />

Employed<br />

(excluding<br />

public<br />

servants)<br />

0.4413 0.3501 6.06 0.3840 0.4864<br />

Executive<br />

employee<br />

0.0284 0.0257 0.54 0.0267 0.1613<br />

Housewife 0.0665 0.1034 4.22 0.0897 0.2858<br />

Retired 0.1577 0.2205 5.05 0.1972 0.3979<br />

18 6 Age 6 30 0.1203 0.0920 2.92 0.1020 0.3033<br />

30 < Age 6 40 0.1644 0.0941 6.88 0.1203 0.3253<br />

40 < Age 6 50 0.1922 0.1385 4.80 0.1585 0.3652<br />

50 < Age 6 60 0.1753 0.1736 0.22 0.1742 0.3793<br />

Age > 60 0.3476 0.5016 10.05 0.4443 0.4969<br />

Log account<br />

volume in<br />

2003<br />

9.0002 9.7146 10.63 9.4489 2.1695<br />

Observations 1648 2792 4440 4440<br />

The t-test refers to a test of the null hypothesis that the mean of the sample with<br />

self-managed accounts equals the mean of the sample with accounts run by a bank<br />

financial advisor (BFA).<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 513<br />

January 2003 to October 2005. For 4447 of those, we have detailed<br />

information on whether particular trades were executed following<br />

consultation with a bank financial advisor (a bank employee) or<br />

without such consultation. Accordingly, we construct a dummy<br />

variable for bank financial advisor use (BFA) that takes the value<br />

of 1 if the customer has consulted with a BFA at least once during<br />

the observation period and 0 if the customer never consulted a BFA<br />

during the period. For comparability’s sake, we match these accounts<br />

to the same regions as in the brokerage data set and use<br />

the same regional variables and (virtually) the same set of account<br />

owner characteristics as for the brokerage sample.<br />

Table 2 presents descriptive statistics for the bank sample.<br />

Again we distinguish between self-managed accounts and accounts<br />

that are at least partly managed by an advisor. Male account<br />

owners are in a minority in this sample (46.3%) and they are underrepresented<br />

(again as indicated by the corresponding t-statistics in<br />

the table) among those customers who consulted a bank financial<br />

advisor before executing some trade(s). As expected, retirees and<br />

housewives are much more strongly represented in the bank sample<br />

than in the brokerage sample. They comprise just fewer than<br />

30% of the observations and they are overrepresented among advised<br />

customers. The majority of account owners are at least<br />

50 years old, and those above 60 are overrepresented among advised<br />

customers. The average account volume at the start of<br />

2003 was slightly higher in this sample, namely 12,694 Euro. The<br />

average monthly turnover rate was 5.8%, somewhat higher than<br />

in the brokerage sample, and smaller for advised customers than<br />

for those who never consulted a BFA.<br />

Finally, Table 2 reports performance figures for the bank accounts,<br />

showing that raw returns and Sharpe ratios are on average<br />

significantly lower for BFA-assisted customers than for self-directed<br />

customers. Given the different composition and advisory models<br />

of the two samples, it will be interesting to see if findings on the<br />

contribution of financial advice to account performance persist (or<br />

differ) across samples.<br />

3.3. Measuring account performance<br />

In this paper we are interested in the effect of financial advice<br />

on portfolio performance and portfolio risk and in particular on<br />

abnormal returns. In order to compute monthly portfolio returns,<br />

we assume as in Dietz (1968) that all transactions occur in the<br />

middle of a given month:<br />

VE VB CF<br />

R ¼<br />

VB þ 0:5 CF<br />

where VE is the value of the portfolio at end of month including<br />

earned dividends and coupons, VB the market value of the portfolio<br />

at beginning of month, and CF is the net cash flow for month t from<br />

purchases (enter positively) and sales of securities(enter negatively)<br />

at transaction prices<br />

Monthly returns from (1) are winsorized by treating returns<br />

that fall into the first or the 100th percentile as missing values. 7<br />

We construct log returns and use them and the standard regression<br />

model in (2) to estimate abnormal (log) returns for each portfolio<br />

based on CAPM.<br />

rp;t rf ;t ¼ ap þ b pðrM;t rf ;tÞþep;t ð2Þ<br />

where ap is the estimated abnormal return (Jensen’s Alpha) for portfolio<br />

p, b p is estimated market beta for portfolio p, r M,t is log return<br />

7 Extreme monthly return observations were treated as missing (and not set to the<br />

upper/lower boundary that would be customary for winsorization) because (a) they<br />

most likely represent erroneous data, and (b) we do not lose customers but just single<br />

months. As a consequence, some customers have only 33 instead of 34 monthly<br />

return observations.<br />

ð1Þ


514 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

of the Euro-denominated MSCI-World Index in month t, rf,t is log return<br />

on the 1-month Euribor, e p,t is the error term of regression for<br />

portfolio p.<br />

In order to test robustness of our results to the way abnormal<br />

returns are computed, we also present results for an alternative<br />

estimate of excess returns based on a four-factor model proposed<br />

by Carhart (1997) to measure portfolio performance. The model<br />

is specified as follows:<br />

ri;t ¼ ai þ b 1iRm;t þ b 2iSMBt þ b 3iHMLt þ b 4iMOMt þ eit<br />

where the intercept ai measures risk-adjusted monthly abnormal<br />

portfolio returns, ri;t denotes monthly excess returns on portfolio i<br />

relative to the risk-free rate which is captured by monthly returns<br />

on the JP Morgan 3 Month Euro Cash Index, R m,t denotes the excess<br />

return on the market portfolio which we approximate by the comprehensive<br />

German CDAX Performance Index, SMB t, HML t, and<br />

MOMt correspond to monthly returns on size, value premium and<br />

momentum portfolios. The size portfolio return (SMB) is approximated<br />

by the difference in monthly returns on the small cap SDAX<br />

index and the large cap DAX 30 index. The book-to-market portfolio<br />

return (HML) is approximated by the return difference between the<br />

MSCI Germany Value Index and the MSCI Germany Growth Index.<br />

Finally, the momentum portfolio return (MOM) is the difference in<br />

monthly returns between a group of stocks with recent above-average<br />

returns and another group of stocks with recent below-average<br />

returns. The group with above-average returns is defined as the top<br />

30% of stocks from the CDAX index over the past 11 months and the<br />

below-average group contains the lowest 30% of stocks from the<br />

same index over the same time period.<br />

4. Performance record of financial advisors<br />

For many brokerage clients, a natural first step towards deciding<br />

whether to use an IFA or not would be to compare the historical performance<br />

of accounts run with IFA involvement and those run without<br />

it. Similarly, bank customers would like to know if those who have<br />

contacted bank financial advisors have done better on average than<br />

those who did not. Even in the absence of official records (indeed neither<br />

the broker or the bank compute or print portfolio performance<br />

0 .05 .1 .15<br />

ð3Þ<br />

records for their clients), prospective clients may still be influenced<br />

by the experiences of existing clients through word of mouth.<br />

Fig. 1 plots histograms of average monthly log returns over our<br />

observation period for brokerage accounts that were self-managed<br />

and for those run with IFA input. Self-managed accounts exhibit a<br />

more symmetric distribution, while advised accounts show higher<br />

mass at the lower end of returns. Table 1 shows monthly logarithmic<br />

returns. The sample mean log monthly return on IFA accounts<br />

over this period is actually lower than that of self-managed accounts:<br />

0.63% versus 1.01%. This corresponds to a difference in annual<br />

rates of return of 5% points (7.9% for advised customers versus<br />

12.9% for those who invested alone). Even though the brokerage<br />

house itself neither collected nor published such statistics, the difference<br />

seems rather hard to miss. Table 1 confirms that IFA accounts<br />

are also characterized by lower abnormal returns than<br />

self-managed accounts, regardless of whether we use a single-factor<br />

model based on the MSCI-World Index or whether we use a<br />

four-factor model based on German stock data.<br />

These lower returns offered by IFAs are combined with lower<br />

average variance of portfolio returns raising the possibility that<br />

they simply reflect an efficient risk-return tradeoff. Strikingly,<br />

however, the sample average of the Sharpe ratio on advised accounts<br />

is also lower than that on self-run brokerage accounts, suggesting<br />

that advisees ‘paid’ on average a higher cost (in terms of<br />

returns) to attain lower risk than what was available to self-managed<br />

accounts. Fig. 2 shows that the distribution of total portfolio<br />

variance under IFAs is ‘squeezed’ towards values closer to zero<br />

compared to what is produced by individuals managing their accounts,<br />

but Fig. 3 shows a much greater heterogeneity in Sharpe ratios<br />

among advised customers than among the rest.<br />

Comparison of IFA and non-IFA accounts also shows a rather<br />

small difference in frequency of trades across the two types of accounts,<br />

but a much more pronounced one when average portfolio<br />

turnover (which is sensitive to the size of purchases) is considered:<br />

the average turnover rate is more than double for IFA accounts.<br />

Looking at Figs. 4 and 5, both measures tend to be clustered closer<br />

to zero for self-managed accounts. In other words, IFAs get commission<br />

based on the volume of purchases and tend to exhibit<br />

greater purchases than individual clients on average. IFA accounts<br />

Self-managed Run by Financial Advisor<br />

-.03 -.02 -.01 0 .01 .02 .03 -.03 -.02 -.01 0 .01 .02 .03<br />

Monthly Returns<br />

Fig. 1. The distributions of log monthly returns.


0 .1 .2<br />

0 .05 .1<br />

0 .005 .01 .015 0 .005 .01 .015<br />

Variance of Monthly Returns<br />

tend also to be larger, and are therefore associated with larger positions<br />

and trades.<br />

Finally, IFA accounts tend to exhibit far greater diversification<br />

than those run by individuals alone. The average share of directly<br />

held stocks among self-managed accounts is just under 60%, while<br />

that for IFA accounts is about 20%. This seems consistent with<br />

incentives to sell mutual funds that IFAs have.<br />

Table 2 presents a similar comparison for bank customers who<br />

have used the advice of bank employees prior to making trades versus<br />

those who have not. Accounts of customers who have resorted to<br />

bank advisors exhibit on average lower returns, comparable variance,<br />

much lower Sharpe ratios, and smaller shares of directly held<br />

stocks than those who did not approach the benchmark. Unlike what<br />

we found for brokerage clients, turnover rates of those who made<br />

use of bank advice were on average lower than of those who did not.<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 515<br />

Self-managed Run by Financial Advisor<br />

Fig. 2. The distributions of the variance of monthly returns.<br />

Self-managed Run by Financial Advisor<br />

-.2 0 .2 .4 .6 .8 -.2 0 .2 .4 .6 .8<br />

Sharpe ratio<br />

Fig. 3. The distribution of the Sharpe ratio.<br />

All in all, performance records of IFAs and BFAs during this sample<br />

period do not appear favorable towards advised accounts, especially<br />

in terms of the risk-return tradeoff offered. The deeper<br />

question is, of course, whether these differences are due to financial<br />

advisors themselves or to the customers they tend to attract.<br />

It is to this household finance question that we now turn, focusing<br />

first on IFAs and then on BFAs.<br />

5. Who has a financial advisor?<br />

We first consider which client characteristics of the brokerage<br />

firm or bank contribute to the client’s account being run with<br />

advisor input. A priori, it may be that advisors tend to be matched<br />

with younger, less experienced and less wealthy investors, who


516 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

Fraction<br />

0 .2 .4 .6 .8<br />

Fraction<br />

0 .5<br />

0 .5 1 1.5 2 0 .5 1 1.5 2<br />

Number of Trades per Year / '000 of Account Volume<br />

need them most; or that they are matched with older, more experienced<br />

and wealthier investors who can pay them most.<br />

Table 3 reports linear probability regressions of whether the<br />

client makes use of an IFA or a BFA, respectively. 8 We control for<br />

time-invariant characteristics (such as availability and cost of financial<br />

advice and characteristics of investor pools) in the region<br />

(columns 1 and 3) and zip code (columns 2 and 4). 9 We see that,<br />

8 Results from probit models (omitting zip-code dummies) deliver very similar<br />

results to the linear probability models, and are available on request.<br />

9 The German Zip Code (Postleitzahlen) is a five digit number consisting of the wider<br />

area that is placed on the thousandth position and the postal district (the unit, tenth<br />

and hundredth positions). Regressions in columns 1, 3, 4 and 6 include dummies for<br />

broader regions (<strong>Dr</strong>esden, Berlin, Hamburg, Hannover, Dusseldorf, Bonn, Frankfurt,<br />

Stuttgart, Munich and Nuremberg), while regressions in columns 2 and 5 include<br />

dummies at the zip code level. There are 5652 zip code dummies for the brokerage<br />

sample and 646 for the bank sample.<br />

Self-managed Run by Financial Advisor<br />

Fig. 4. The distribution of number of trades (per ‘000 account volume).<br />

Self-managed Run by Financial Advisor<br />

0 .1 .2 .3 .4 .5 .6 0 .1 .2 .3 .4 .5 .6<br />

Turnover Rate<br />

Fig. 5. The distribution of the monthly turnover rate.<br />

given other characteristics, males are less likely to use an advisor,<br />

consistent with the view that males tend to have more (over)confidence.<br />

Older clients (over 50) have a significantly greater probability<br />

than investors between 18 and 30 of using an advisor, by about 10%<br />

points in both samples. Wealthier brokerage clients, as proxied by<br />

the beginning-of-period account size to minimize endogeneity problems,<br />

are significantly more likely to use IFA or BFA.<br />

Married clients are less likely to use an IFA, controlling for other<br />

factors, probably because spouses can be used as sounding boards.<br />

An extra year of self-reported experience with the relevant financial<br />

products increases the probability of using an IFA. We do not<br />

have information on the marital status and trading experience of<br />

bank customers, but their professional status is statistically<br />

insignificant.<br />

Overall our regressions show that advisors are more likely to be<br />

matched with wealthier (as measured by account volume), older,


Table 3<br />

The determinants of having the account run by an IFA or a BFA.<br />

Brokerage sample (IFAs) Bank sample(BFAs)<br />

(1) (2) (3) (4)<br />

Male 0.062 ***<br />

0.061 ***<br />

0.049 ***<br />

0.038 **<br />

(12.06) (10.92) (3.14) (2.15)<br />

Employee 0.057 **<br />

0.061 **<br />

0.054 ***<br />

0.039 *<br />

(2.44) (2.52) (2.93) (1.91)<br />

30 < Age 6 40 0.031 ***<br />

0.025 **<br />

0.050 0.060<br />

(3.32) (2.40) (1.44) (1.49)<br />

40 < Age 6 50 0.004 0.004 0.013 0.021<br />

(0.36) (0.35) (0.38) (0.57)<br />

50 < Age 6 60 0.023 **<br />

0.021 *<br />

0.053 *<br />

0.042<br />

(2.20) (1.81) (1.67) (1.16)<br />

Age > 60 0.088 ***<br />

0.088 ***<br />

0.112 ***<br />

0.089 ***<br />

(7.53) (6.71) (3.87) (2.73)<br />

Log account volume 0.045 ***<br />

0.044 ***<br />

0.027 ***<br />

0.034 ***<br />

(22.07) (20.19) (7.51) (7.92)<br />

Self-employed 0.060 **<br />

0.064 **<br />

(2.48) (2.55)<br />

Experience/100 0.159 ***<br />

0.148 ***<br />

(3.45) (3.03)<br />

Married 0.023 ***<br />

0.025 ***<br />

(5.61) (5.10)<br />

Executive 0.003 0.032<br />

(0.06) (0.59)<br />

Housewife 0.002 0.016<br />

(0.06) (0.49)<br />

Retired 0.025 0.009<br />

(1.11) (0.34)<br />

Constant 0.358 ***<br />

0.316 ***<br />

0.366 ***<br />

0.303 ***<br />

(12.58) (10.34) (7.68) (6.58)<br />

Observations 28,321 28,321 4440 4440<br />

R-squared 0.09 0.37 0.05 0.23<br />

Zip code dummies No Yes No Yes<br />

The table reports estimates from a linear probability model of having an Independent<br />

Financial Advisor or a bank financial advisor. Log account volume is measured<br />

in January 2003. All regressions include regional dummies (absorbed by zip code<br />

dummies where present). Asymptotic standard errors corrected for clustering at the<br />

zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

more experienced, single, and female investors. Such investors<br />

have better reasons to want to delegate to advisors, such as high<br />

opportunity cost or low inclination to spend time managing investments,<br />

as well as sizeable wealth. The results are remarkably consistent<br />

across the two samples, and robust to inclusion of zip code<br />

dummies that control for unobserved factors at the local level.<br />

Since IFAs and BFAs earn more on wealthy clients with high opportunity<br />

costs of time, they seem to go for the big players who have a<br />

lot to invest, rather than for the younger, smaller, inexperienced<br />

investors who have a lot to learn.<br />

6. Independent financial advisors and portfolio performance<br />

We now turn to how IFA use affects account performance once<br />

we control for client characteristics. An important estimation issue<br />

is omitted variable bias: unobserved factors may simultaneously<br />

affect the probability of using an advisor as well as account performance.<br />

For instance, our data do not report willingness to undertake<br />

financial risk: more risk averse clients may be inclined to consult an<br />

advisor and to invest in a safer portfolio, thus influencing account<br />

returns and variance. Other factors could also influence both advisor<br />

use and account performance: financial literacy and sophistication;<br />

attitudes developed in formative years, e.g. through parental influence<br />

or observation of others in the parental social circle; social attitudes,<br />

such as trust in others, that have been shown to influence<br />

portfolio composition (e.g. participation in stockholding) and dele-<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 517<br />

gation to a financial advisor. In order to attenuate this problem, we<br />

control for as many possible factors that are observed in our data<br />

set as well as for regional dummies and a finer classification of zip<br />

code dummies.<br />

A second issue is the potential endogeneity of the choice of consulting<br />

a financial advisor. Investors with low performing portfolios<br />

may be induced to use a financial advisor by the media or<br />

specific campaigns that advisor-assisted portfolios perform better.<br />

A negative correlation between advisor consultation and, say, returns,<br />

might therefore be driven by the effectiveness of these campaigns,<br />

rather than by a negative role of financial advisors per se. In<br />

order to handle this possibility, we present (in Appendix A) instrumental<br />

variable estimates using as instrument the local GDP share<br />

of financial services. Results are consistent with our OLS findings<br />

below.<br />

6.1. IFA effect on portfolio returns<br />

Table 4 presents OLS estimates regarding the influence of IFA<br />

use on raw net returns, and on abnormal net returns, constructed<br />

on the basis of a single-factor and of a four-factor model in the<br />

spirit of Carhart (1997). Columns 1 and 2 report estimated effects<br />

on average portfolio returns, controlling for investor characteristics<br />

and regional dummies. Model (2) adds zip code dummies as controls<br />

for time-invariant local characteristics. IFA effects are almost<br />

identical in both models: negative and statistically significant at<br />

the 1% level, implying that IFA use reduces monthly log returns<br />

by roughly 0.4% points. Thus, the lower returns for advised accounts<br />

in descriptive statistics survive controls for personal and regional<br />

characteristics.<br />

Even if IFAs reduce raw returns, they might still be found to create<br />

value by increasing risk adjusted returns. Columns 3 and 4 in<br />

Table 4 report OLS regressions for alphas from a model with the return<br />

on the MSCI world index as the single factor (denoted Jensen’s<br />

alpha). The IFA contribution is again negative and of similar magnitude<br />

as for raw returns, once characteristics of the account owner<br />

and region are taken into account. In columns 5 and 6, we examine<br />

robustness with respect to using the four-factor model for German<br />

stock markets outlined above. The strongly statistically significant<br />

negative effect of IFA use is observed regardless of whether we use<br />

a single or a four-factor model, and its size is remarkably similar<br />

with the other models and with the descriptive statistics from Table<br />

1.<br />

Across all models, male gender is found to detract from account<br />

returns, consistent with the literature on overconfidence. Years of<br />

experience tend to contribute to higher total return, albeit by a<br />

small estimated amount. This is consistent with recent studies<br />

indicating that the magnitude of investment mistakes decreases<br />

with sophistication and experience. 10<br />

Findings in this section imply that involvement of IFAs with<br />

brokerage accounts tends to reduce both raw and abnormal returns,<br />

even after investor and area characteristics are taken into account.<br />

Our results are consistent with the cost of financial advice<br />

exceeding, on average, any benefits from informational contributions.<br />

Importantly, this does not necessarily imply that (some) IFAs<br />

engage in misselling or that all IFAs give uniformly bad advice.<br />

10 For example, Feng and Seasholes (2005) ask whether investor sophistication and<br />

trading experience eliminate behavioral biases, such as the disposition effect, using<br />

data from the PR of China. They proxy sophistication mainly by the number of trading<br />

rights (indicating the number of methods to trade) and an indicator of initial portfolio<br />

diversification, both at the start of the observation period. Experience is proxied by the<br />

number of positions taken by investor i up until date t, a time-varying covariate. They<br />

conclude that sophistication and experience eliminate the reluctance to realize losses,<br />

but only reduce the propensity to realize gains by 37%. See also Grinblatt and<br />

Keloharju (2001), Feng and Seasholes (2005), and Lusardi and Mitchell (2007).


518 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

Table 4<br />

The determinants of portfolio returns in the brokerage sample.<br />

Log returns Jensen’s alpha Alpha 4 factor model<br />

(1) (2) (3) (4) (5) (6)<br />

Financial advisor (IFA) 0.004 ***<br />

0.004 ***<br />

0.004 ***<br />

0.004 ***<br />

0.004 ***<br />

0.004 ***<br />

(26.45) (18.33) (26.88) (18.22) (27.03) (17.11)<br />

Male 0.001 ***<br />

0.000 ***<br />

0.000 ***<br />

0.000 ***<br />

0.001 ***<br />

0.001 ***<br />

(4.34) (3.14) (3.95) (2.99) (5.37) (4.17)<br />

Married 0.000 0.000 0.000 0.000 0.000 **<br />

0.000<br />

(0.79) (0.22) (1.16) (0.43) (2.12) (1.49)<br />

Employee 0.001 0.001 0.001 0.001 0.001 *<br />

0.001 *<br />

(1.57) (1.43) (1.47) (1.39) (1.95) (1.72)<br />

Self-employed 0.001 *<br />

0.001 *<br />

0.001 *<br />

0.001 *<br />

0.001 **<br />

0.002 **<br />

(1.91) (1.77) (1.86) (1.74) (2.35) (2.11)<br />

Experience/100 0.002 **<br />

0.001 0.001 0.000 0.001 0.001<br />

(2.28) (0.98) (1.58) (0.43) (1.31) (0.53)<br />

30 < Age 6 40 0.000 *<br />

0.001 *<br />

0.000 *<br />

0.001 *<br />

0.000 0.000<br />

(1.84) (1.70) (1.86) (1.71) (0.89) (1.07)<br />

40 < Age 6 50 0.000 0.000 0.000 0.000 0.000 0.000<br />

(0.32) (0.37) (0.46) (0.43) (1.36) (0.97)<br />

50 < Age 6 60 0.000 0.000 0.000 0.000 0.000 0.000<br />

(0.15) (0.43) (0.18) (0.48) (1.11) (0.33)<br />

Age > 60 0.001 *<br />

0.000 0.000 0.000 0.001 ***<br />

0.001<br />

(1.80) (0.95) (1.64) (0.79) (2.63) (1.58)<br />

Log account volume 0.000 ***<br />

0.000 ***<br />

0.000 ***<br />

0.000 ***<br />

0.000 ***<br />

0.000 ***<br />

(6.92) (4.81) (6.68) (4.62) (7.51) (5.24)<br />

Constant 0.008 ***<br />

0.008 ***<br />

0.008 ***<br />

0.008 ***<br />

0.007 ***<br />

0.008 ***<br />

(11.92) (9.78) (11.24) (9.36) (9.48) (7.90)<br />

Observations 28,321 28,321 28,321 28,321 28,321 28,321<br />

R-squared 0.03 0.23 0.02 0.22 0.02 0.22<br />

Zip code dummies No Yes No Yes No Yes<br />

Log account volume is measured in January 2003. All regressions include regional dummies (absorbed by zip code dummies where present). Asymptotic t-statistics corrected<br />

for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

6.2. IFA effect on variance of returns and Sharpe ratio<br />

The finding that IFAs tend to lower both raw and abnormal account<br />

returns, given investor characteristics, need not be negative,<br />

if IFAs ensure that clients are exposed to smaller portfolio risk.<br />

Descriptive statistics above seem to be pointing in this direction.<br />

We therefore turn next to the effect of IFA involvement on variance<br />

of monthly portfolio returns and on the risk-return tradeoff as captured<br />

by the Sharpe ratio. Table 5 reports our findings.<br />

Column 1 reports OLS results for a model with regional dummies,<br />

whereas column 2 reports results when zip-code dummies<br />

are included. In both models, IFAs reduce portfolio risk in line with<br />

descriptive results. Being male, inexperienced, single, self-employed,<br />

and with a smaller account all contribute significantly to<br />

higher total portfolio risk. Results on control variables are intuitive.<br />

For example, larger accounts should allow more diversification,<br />

and we indeed find below that they tend to have smaller portfolio<br />

shares in directly held stocks.<br />

Set against the negative effects of IFAs on account returns, their<br />

moderating effect on variance raises the question of whether IFAs<br />

help achieve an efficient risk-return tradeoff. We present two<br />

regressions (columns 3 and 4), one with regional and the other<br />

with zip code dummies. Both show a statistically significant negative<br />

effect of IFAs on Sharpe ratios. Male gender contributes to inferior<br />

risk-return tradeoffs, consistent with overconfidence; while<br />

more experienced or married investors tend to achieve better<br />

tradeoffs. Interestingly, self-employed and older clients are seen<br />

to have a tendency to expose themselves to more risk than what<br />

is efficient for a given increase in expected return. Finally, wealthier<br />

investors tend to achieve better risk-return tradeoffs, presumably<br />

by exploiting economies of scale in asset management. We<br />

conclude from Table 5 that IFAs tend to reduce portfolio risk but<br />

do not compensate sufficiently for lower returns: IFA use decreases<br />

ex post portfolio efficiency.<br />

6.3. IFA effect on trading, turnover, and diversification<br />

What type of behavior underlies our results on returns and risk?<br />

The fact that IFAs earn commissions mainly when the account<br />

owner purchases mutual funds creates an incentive for them to<br />

encourage fund purchases. The first two columns of Table 6 examine<br />

the effect of IFA on the number of purchases per month scaled<br />

by account volume. These exclude account transactions from corporate<br />

actions, periodic saving plan investments and portfolio<br />

transfers, so as to be more directly linked to the IFA incentives to<br />

sell specific financial instruments.<br />

Our results imply a negative effect of IFAs on the standardized<br />

number of purchases. Purchases result in transactions costs and<br />

could contribute to lower net returns, but it appears that the negative<br />

effect of IFAs on net returns reported above does not result<br />

simply from an increased frequency of purchases. The regression<br />

does confirm the positive role of male gender found in other studies<br />

(see above). Financial experience is estimated to reduce the<br />

number of purchases, consistent with Dorn and Huberman<br />

(2005) finding that respondents with longer investment experience<br />

trade less, but the effect is not statistically significant. Account<br />

holders between 40 and 60 are significantly more likely to<br />

engage in purchases than other age groups. Subject to the<br />

provision interpreting age effects, this finding is consistent with<br />

them being in the asset accumulation phase, prior to entering<br />

retirement.<br />

Although we do not find a simple channel through frequency of<br />

trading, this does not mean that IFAs do not respond to incentives<br />

offered by commissions. It is useful to recall that commissions are


Table 5<br />

The determinants of portfolio return variance and Sharpe ratios in the brokerage sample.<br />

Variance of monthly returns Sharpe ratio<br />

(1) (2) (3) (4)<br />

Financial advisor (IFAs) 0.001 ***<br />

0.001 ***<br />

0.056 ***<br />

0.052 ***<br />

(14.74) (9.95) (13.87) (10.92)<br />

Male 0.001 ***<br />

0.001 ***<br />

0.022 ***<br />

0.021 ***<br />

(15.82) (13.11) (10.31) (7.85)<br />

Married 0.000 ***<br />

0.000 ***<br />

0.006 ***<br />

0.006 ***<br />

(8.01) (6.27) (3.26) (2.62)<br />

Employee 0.000 *<br />

0.000 0.021 **<br />

0.017<br />

(1.77) (0.24) (2.10) (1.39)<br />

Self-employed 0.001 ***<br />

0.001 ***<br />

0.033 ***<br />

0.029 **<br />

(5.18) (2.93) (3.22) (2.21)<br />

Experience/100 0.000 0.000 0.056 ***<br />

0.043 **<br />

(0.50) (0.47) (3.74) (2.31)<br />

30 < Age 6 40 0.000 *<br />

0.000 0.001 0.002<br />

(1.77) (1.19) (0.17) (0.42)<br />

40 < Age 6 50 0.000 ***<br />

0.000 ***<br />

0.012 ***<br />

0.013 **<br />

(5.46) (3.99) (2.59) (2.33)<br />

50 < Age 6 60 0.000 ***<br />

0.000 ***<br />

0.014 ***<br />

0.013 **<br />

(5.84) (4.21) (2.77) (2.19)<br />

Age > 60 0.001 ***<br />

0.000 ***<br />

0.018 ***<br />

0.015 **<br />

(6.22) (3.99) (3.58) (2.45)<br />

Log account volume 0.001 ***<br />

0.001 ***<br />

0.023 ***<br />

0.021 ***<br />

(39.58) (31.94) (26.96) (20.47)<br />

Constant 0.007 ***<br />

0.008 ***<br />

0.066 ***<br />

0.064 ***<br />

(38.35) (32.36) (5.22) (4.12)<br />

Observations 28,321 28,321 28,321 28,321<br />

R-squared 0.12 0.31 0.05 0.25<br />

Zip code dummies No Yes No Yes<br />

Log account volume is measured in January 2003. All regressions include regional dummies (absorbed by zip code dummies where present).<br />

Asymptotic t-statistics corrected for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

linked to the size, and not merely to the frequency of purchases.<br />

The third and fourth columns of Table 6 show a positive and<br />

strongly statistically significant effect of IFAs on average account<br />

turnover. This could be part of the explanation for why IFAs contribute<br />

negatively to portfolio returns. 11 Again, males are more<br />

likely to have larger account turnover and more experienced investors<br />

are less likely to turn over their portfolio frequently. Younger<br />

investors, between 30 and 60 years of age, are estimated to have<br />

higher purchase turnovers, as they actively expand their portfolios.<br />

A different perspective on the role of IFAs applies to encouraging<br />

diversification. Columns 5 and 6 of Table 6 report a negative IFA<br />

effect on the average share of directly held stocks in the account,<br />

even after characteristics of account holders and areas are controlled<br />

for. 12 This finding is consistent both with the descriptive statistics<br />

at the start of the paper and the incentive of IFAs to sell<br />

mutual funds. It is also one channel through which the reduction<br />

in portfolio variance that we found in Table 5 is likely to be accomplished<br />

by IFAs.<br />

Controlling for other factors, males tend to put larger shares of<br />

their account in directly held stocks, suggesting overconfidence in<br />

portfolio behavior, in addition to the gender effects on frequency of<br />

trading and on the volume of purchases. 13 Interestingly, experience<br />

tends to lower the share of directly held stocks, dampening overcon-<br />

11 Higher turnover might be motivated simply by commissions but also by an<br />

incentive of IFAs to justify their fees by rebalancing client portfolios (see e.g.<br />

Lakonishok et al., 1992).<br />

12 Since the share of single stocks is bound between zero and one, we also run probit<br />

regressions. Results deliver very similar results to the OLS estimates, and are available<br />

on request.<br />

13 Being married tends to have the opposite effect, presumably because more people<br />

are at risk and maybe vocal in encouraging diversification. Employees and selfemployed<br />

account owners tend to invest more in directly held stocks, probably<br />

because of their increased social interactions.<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 519<br />

fidence rather than encouraging account owners to manage direct<br />

investments in stocks. The conclusion from the regression analysis<br />

is that IFAs seem to boost the volume of purchases, while reducing<br />

the fraction of the account invested in directly held stocks.<br />

7. Bank financial advisors and portfolio performance<br />

Given the rather striking nature of the estimated contribution of<br />

IFAs to portfolio performance, the question arises as to whether<br />

our results are specific to brokerage accounts, e.g. because of selectivity<br />

into these types of accounts, or because financial advisors are<br />

independent and not accountable to the financial institution. For<br />

this reason, we consider a second data set of investment accounts,<br />

this time from a large German commercial bank.<br />

Our discussions with the brokerage and the commercial bank<br />

suggest that there are important similarities and differences between<br />

incentives facing IFAs and BFAs. For example, upfront loads<br />

for mutual funds, a key component of any incentive scheme, are<br />

typically fixed by the mutual fund producer and therefore identical<br />

for all sales organizations. Although the bank does not funnel all<br />

commissions through to its BFAs, it gives powerful non-monetary<br />

incentives to its sales force through its sales control system. 14 On<br />

the other hand, IFAs are not subject to the constraints imposed by<br />

banks on BFAs. In fact, many banks not only narrow down the menu<br />

of financial products offered to investors, but also provide extra<br />

incentives for their agents to advise clients to purchase funds or<br />

structured products produced by the bank itself or by one of its<br />

subsidiaries. 15 We expect the negative association between BFA<br />

14<br />

Another similarity refers to legal fines for any detected misselling. Since they are<br />

a function of the loss to the client they should be identical across IFAs and banks.<br />

15<br />

Yoong and Hung (2009) extend Ottaviani and Inderst (2009) to address this kind<br />

of self-dealing.


520 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

Table 6<br />

The determinants of trading, turnover and diversification in the brokerage sample.<br />

Number of trades Turnover rate Share of single stocks<br />

(1) (2) (3) (4) (5) (6)<br />

Financial advisor (IFA) 0.008 ***<br />

0.009 **<br />

0.062 ***<br />

0.053 ***<br />

0.353 ***<br />

0.341 ***<br />

(2.59) (2.14) (16.77) (15.21) (47.19) (37.72)<br />

Male 0.035 ***<br />

0.036 ***<br />

0.018 ***<br />

0.017 ***<br />

0.077 ***<br />

0.079 ***<br />

(16.04) (12.32) (16.10) (13.15) (14.71) (12.37)<br />

Married 0.001 0.001 0.001 0.000 0.027 ***<br />

0.029 ***<br />

(0.32) (0.19) (1.37) (0.14) (6.03) (5.03)<br />

Employee 0.010 0.015 0.001 0.001 0.082 ***<br />

0.068 **<br />

(0.80) (0.89) (0.19) (0.26) (2.94) (1.99)<br />

Self-employed 0.006 0.012 0.004 0.005 0.128 ***<br />

0.108 ***<br />

(0.46) (0.71) (0.99) (0.91) (4.47) (3.10)<br />

Experience/100 0.020 0.003 0.073 ***<br />

0.054 ***<br />

0.571 ***<br />

0.556 ***<br />

(1.07) (0.12) (8.58) (5.42) (15.22) (11.91)<br />

30 < Age 6 40 0.005 0.004 0.004 **<br />

0.003 0.000 0.001<br />

(1.39) (0.80) (2.05) (1.17) (0.03) (0.04)<br />

40 < Age 6 50 0.015 ***<br />

0.016 ***<br />

0.010 ***<br />

0.009 ***<br />

0.028 **<br />

0.026 *<br />

(4.02) (2.95) (4.29) (3.44) (2.50) (1.88)<br />

50 < Age 6 60 0.024 ***<br />

0.026 ***<br />

0.016 ***<br />

0.014 ***<br />

0.061 ***<br />

0.064 ***<br />

(5.35) (4.28) (6.53) (4.78) (5.06) (4.33)<br />

Age > 60 0.020 ***<br />

0.021 ***<br />

0.011 ***<br />

0.011 ***<br />

0.077 ***<br />

0.072 ***<br />

(4.11) (3.17) (4.28) (3.50) (6.27) (4.74)<br />

Log account volume 0.015 ***<br />

0.016 ***<br />

0.009 ***<br />

0.007 ***<br />

0.027 ***<br />

0.030 ***<br />

(13.23) (10.75) (15.95) (12.03) (17.05) (14.69)<br />

Constant 0.097 ***<br />

0.093 ***<br />

0.101 ***<br />

0.089 ***<br />

0.689 ***<br />

0.735 ***<br />

(6.21) (4.57) (16.08) (12.43) (21.07) (18.62)<br />

Observations 28,303 28,303 28,321 28,321 28,321 28,321<br />

R-squared 0.02 0.21 0.07 0.32 0.14 0.32<br />

Zip code dummies No Yes No Yes No Yes<br />

Number of trades is expressed as a fraction of account volume in ‘000. Log account volume is measured in January 2003. All regressions include regional dummies (absorbed<br />

by zip code dummies where present). Asymptotic t-statistics corrected for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

Table 7<br />

The determinants of portfolio returns, variance and Sharpe ratio in the bank sample.<br />

Log monthly returns Variance of monthly returns Sharpe ratio<br />

(1) (2) (3) (4) (5) (6)<br />

Financial advisor (BFA) 0.003 ***<br />

0.003 ***<br />

0.001 ***<br />

0.001 *<br />

0.180 ***<br />

0.178 ***<br />

(9.20) (8.07) (2.80) (1.92) (12.20) (10.20)<br />

Male 0.001 **<br />

0.001 *<br />

0.001 ***<br />

0.001 ***<br />

0.031 **<br />

0.033 **<br />

(2.18) (1.75) (3.05) (2.63) (2.37) (2.13)<br />

Employee 0.001 ***<br />

0.001 *<br />

0.001 0.001 0.008 0.005<br />

(2.59) (1.88) (1.49) (0.92) (0.47) (0.25)<br />

Executive 0.001 0.002 0.002 0.002 0.013 0.030<br />

(1.37) (1.57) (1.46) (1.33) (0.31) (0.65)<br />

Housewife 0.001 0.001 *<br />

0.000 0.000 0.008 0.020<br />

(1.58) (1.70) (0.19) (0.14) (0.34) (0.72)<br />

Retired 0.000 0.000 0.000 0.000 0.010 0.001<br />

(0.27) (0.48) (0.22) (0.58) (0.53) (0.04)<br />

30 < Age 6 40 0.001 *<br />

0.001 0.000 0.000 0.034 0.029<br />

(1.86) (1.49) (0.09) (0.14) (1.21) (0.86)<br />

40 < Age 6 50 0.000 0.000 0.002 *<br />

0.001 0.023 0.010<br />

(0.41) (0.32) (1.87) (1.46) (0.91) (0.33)<br />

50 < Age 6 60 0.000 0.001 0.000 0.001 0.023 0.015<br />

(0.64) (1.08) (0.52) (0.65) (0.81) (0.46)<br />

Age > 60 0.002 **<br />

0.002 **<br />

0.001 0.001 0.028 0.012<br />

(2.58) (2.33) (0.88) (1.21) (1.10) (0.40)<br />

Log account volume 0.001 ***<br />

0.000 ***<br />

0.001 ***<br />

0.001 ***<br />

0.023 ***<br />

0.024 ***<br />

(5.44) (3.81) (8.04) (7.15) (5.08) (4.85)<br />

Constant 0.001 0.003 **<br />

0.013 ***<br />

0.013 ***<br />

0.197 ***<br />

0.210 ***<br />

(1.01) (2.18) (8.47) (8.58) (4.33) (4.48)<br />

Observations 4440 4440 4440 4440 4440 4440<br />

R-squared 0.06 0.20 0.04 0.19 0.05 0.19<br />

Zip code dummies No Yes No Yes No Yes<br />

Log account volume is measured in January 2003. All regressions include regional dummies (absorbed by zip code dummies where present). Asymptotic t-statistics corrected<br />

for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.


Table 8<br />

The determinants of turnover and diversification in the bank sample.<br />

Turnover rate Share of single stocks<br />

(1) (2) (3) (4)<br />

Financial advisor (BFA) 0.099 ***<br />

0.121 ***<br />

0.158 ***<br />

0.148 ***<br />

(6.11) (5.91) (16.31) (13.22)<br />

Male 0.072 ***<br />

0.075 ***<br />

0.056 ***<br />

0.055 ***<br />

(4.72) (4.37) (5.88) (4.85)<br />

Employee 0.041 **<br />

0.019 0.045 ***<br />

0.037 ***<br />

(1.98) (0.82) (3.66) (2.59)<br />

Executive 0.147 **<br />

0.137 **<br />

0.046 0.054 *<br />

(2.43) (2.02) (1.63) (1.72)<br />

Housewife 0.035 0.029 0.010 0.004<br />

(1.07) (0.76) (0.62) (0.19)<br />

Retired 0.038 0.023 0.016 0.019<br />

(1.54) (0.84) (1.16) (1.16)<br />

30 < Age 6 40 0.011 0.046 0.072 ***<br />

0.079 ***<br />

(0.38) (1.32) (3.69) (3.52)<br />

40 < Age 6 50 0.016 0.038 0.088 ***<br />

0.099 ***<br />

(0.56) (1.19) (4.69) (4.45)<br />

50 < Age 6 60 0.051 *<br />

0.069 **<br />

0.053 ***<br />

0.061 ***<br />

(1.87) (2.14) (3.04) (3.02)<br />

Age > 60 0.012 0.016 0.014 0.029<br />

(0.47) (0.55) (0.85) (1.58)<br />

Log account volume 0.065 ***<br />

0.062 ***<br />

0.006 ***<br />

0.008 ***<br />

(14.85) (12.42) (2.70) (2.90)<br />

Constant 0.339 ***<br />

0.368 ***<br />

0.275 ***<br />

0.267 ***<br />

(6.53) (7.07) (10.04) (9.73)<br />

Observations 4440 4440 4440 4440<br />

R-squared 0.10 0.26 0.13 0.26<br />

Zip code dummies No Yes No Yes<br />

Log account volume is measured in January 2003. All regressions include regional<br />

dummies (absorbed by zip code dummies where present). Asymptotic t-statistics<br />

corrected for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

use and portfolio returns to be even stronger than in the brokerage<br />

sample.<br />

As with the brokerage sample, we introduce regional and zip<br />

code dummies to capture unobserved heterogeneity. Columns 1<br />

and 2 in Table 7 present OLS results on raw returns, where use of<br />

a BFA is seen to have a statistically significant negative effect.<br />

According to this model, BFAs reduce monthly log returns by<br />

0.3% points, slightly less than IFAs in Table 4 ( 0.4%). However, unlike<br />

IFAs, who were found to reduce overall portfolio risk, BFAs are<br />

found in columns 3 and 4 of Table 7 to increase total portfolio risk.<br />

Given the negative BFA effect on returns and their positive effect<br />

on total risk, we expect a strong negative effect on Sharpe ratios,<br />

and this is confirmed by the last two columns in Table 7. 16 This<br />

pronounced negative effect of BFAs is consistent with our conjecture<br />

from above and with Inderst and Ottaviani (2009) who posit that<br />

advisory standards should be lower for BFAs than for IFAs because<br />

the latter face no internal agency conflicts with costly monitoring.<br />

Consistent with results on IFAs, males exhibit lower returns and<br />

riskier portfolios. The initial size of the account contributes to<br />

higher returns, in levels or normalized by risk, and to lower portfolio<br />

variance. Columns 1 and 2 of Table 8 point to higher turnover<br />

rates (based on purchases) for BFA accounts. The estimated BFA<br />

coefficients are larger than the corresponding IFA ones (Table 6),<br />

suggesting lower advisory standards for BFAs than for IFAs.<br />

Finally, regressions reported in columns 5 and 6 of Table 8 confirm<br />

that BFAs, as indeed IFAs, tend to push towards investing in<br />

mutual funds, consistent with their compensation incentives. All<br />

in all, our analysis of the bank sample produces remarkably consis-<br />

16 The coefficients of the BFA dummy are negative ( 0.18), with values considerably<br />

larger than the corresponding IFA coefficients in Table 5 ( 0.05 in columns 3 and 4).<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 521<br />

tent results with the brokerage sample and points to systematic<br />

negative effects of financial advisors rather than to statistical flukes<br />

or sample peculiarities.<br />

8. Conclusions<br />

We investigate who tends to use a financial advisor, whether<br />

investors tend to produce better account performance on their<br />

own rather than with the help of financial advisors, whether results<br />

depend on the advisory model (IFA versus BFA), and whether<br />

they can be traced to trading behavior and security choice. We also<br />

examine robustness of our findings with respect to asset pricing<br />

model, dummies for wider regions or zip codes, control variables,<br />

and estimation procedure (OLS versus IV).<br />

Our first data set tracks accounts of a major brokerage firm,<br />

some of which are run with the help of an independent financial<br />

advisor (IFA). Sample statistics and regression analysis show that<br />

advisors tend to be matched with wealthier, older, more experienced,<br />

and female investors rather than with poorer, younger<br />

and inexperienced ones. Our second data set comes from a major<br />

commercial bank with branches throughout the country. We find<br />

that also bank clients who tend to consult a bank employee prior<br />

to executing a trade are older, wealthier and more likely to be<br />

female.<br />

Descriptive statistics as well as regression analysis that controls<br />

for investors’ characteristics and characteristics of the region of the<br />

account paint a very similar picture of the role of IFA and BFAs in<br />

account performance. In both samples, advised accounts offer lower<br />

returns than those run by similar investors without advisor input.<br />

Although IFA use reduces total portfolio risk, it still reduces<br />

ex post Sharpe ratios significantly. BFAs increase portfolio variance,<br />

lowering Sharpe ratios even more.<br />

Trading costs and associated commissions earned by both IFAs<br />

and banks certainly contribute to these outcomes, since we find<br />

that advised accounts feature higher portfolio turnover (though<br />

not necessarily more frequent trading) relative to self-managed accounts.<br />

Consistent with their remuneration incentives, financial<br />

advisors tend to encourage lower account shares in directly held<br />

stocks. Robustness analysis suggests that our results on the negative<br />

role of IFAs are not an artifact of endogeneity between account<br />

performance and advisor use, nor of the way we adjust portfolio<br />

net returns for systematic risk.<br />

Our results provide a new perspective on the role of financial<br />

advisors that might be useful for theoretical and policy analysis<br />

of their conflicting incentives, their likely effects, and the need to<br />

regulate them. Based on our findings, it should not be taken for<br />

granted that financial advisors provide their services to small,<br />

young investors typically identified as in need of investment guidance.<br />

Indeed, the opposite is true both for the broker and for the<br />

bank data we consider. The finding stands to reason: financial<br />

advisors with commission-based incomes naturally prefer to devote<br />

time to customers likely to trade on a bigger scale. However,<br />

it also creates doubts as to how viable financial advice is as a solution<br />

to the problem of limited financial literacy in the population.<br />

In view of the rapidly growing literature on investment mistakes,<br />

providing financial advice to inexperienced, naïve investors could<br />

be an alternative to trying to educate them in financial matters,<br />

but financial advisor incentives and tendencies of inexperienced<br />

clients might result in relatively few matches. Other alternatives,<br />

such as simpler products and carefully designed default options,<br />

may be more promising than currently existing forms of financial<br />

advice in averting negative distributional consequences.<br />

Our findings imply that many financial advisors end up collecting<br />

more in fees and commissions than any monetary value they<br />

add to the account. This raises the further question of whether


522 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

advisors overcharge and should be regulated. While the case for<br />

regulation seems much clearer when advisors are matched with<br />

inexperienced investors, negative effects appear even when the<br />

tendency is for experienced investors to be using an advisor. In<br />

such cases, it may be that investors are inattentive and fail to monitor<br />

the advisors effectively; or that they face high opportunity<br />

costs of running accounts by themselves and are willing to pay a<br />

luxury premium to have their advisors run their accounts. What<br />

distinguishes these two cases is customer awareness of the financial<br />

advisor incentives and effects. Even if regulation is not warranted<br />

in both cases, transparency and information on the role<br />

and outcome of financial advice seem crucial. Moreover, this need<br />

is not limited to naïve, inexperienced customers but extends also<br />

to older, experienced ones. Thus, questionnaires on investor experience,<br />

such as those dictated by MIFID (the EU directive aimed at<br />

increasing financial markets transparency and competition),<br />

should not waive the need for information regarding incentives,<br />

ensuing conflicts of interest, and the outcomes of professional<br />

financial advice in terms of portfolio returns and risks, even for<br />

experienced investors.<br />

Acknowledgements<br />

We are grateful to the following individuals for very helpful<br />

comments and suggestions: the editor, an anonymous referee, Raquel<br />

Carrasco, Guenter Franke, Dimitris Georgarakos, Luigi Guiso,<br />

Jose Martinez, Pedro Mira, Steven Stern, John Rust; participants<br />

in the November 2008 Fundación Ramón Areces conference on<br />

Population Ageing and its Economic Consequences in Madrid, the<br />

2009 SIFR-NETSPAR Pension workshop in Stockholm, the 2009 IM-<br />

AEF conference in Ioannina, the 2009 CSEF-IGIER Symposium in<br />

Economics and Institutions in Capri, the SAVE Conference in Mannheim,<br />

the Bank of Spain 2009 Household <strong>Finance</strong> Conference; and<br />

in the seminar series at the Universities of Exeter, Konstanz, Leicester,<br />

Piraeus, and Central European University in Budapest. We<br />

would like to thank Ralph Bluethgen and Yigitcan Karabulut for<br />

outstanding research assistance and the two financial institutions<br />

for providing us with the administrative data. Haliassos acknowledges<br />

research funding from the German Research Foundation<br />

(Deutsche Forschungsgemeinschaft) under the project Regulating<br />

Retail <strong>Finance</strong>. Jappelli acknowledges funding from the Italian Ministry<br />

of Universities and Research.<br />

Appendix A. Endogeneity of financial advice<br />

Although we define as advised portfolios those portfolios that<br />

are continuously assisted from 2003 to 2006, if performance is persistent<br />

over time one may suspect that portfolio performance actually<br />

induced the choice of the advisor. As mentioned in Section 6,in<br />

such a case, OLS regressions are problematic. In this appendix, we<br />

carry out IV estimation to examine the robustness of our main<br />

findings to possible endogeneity of financial advice. Finding suitable<br />

instruments in our context is not easy, and the robustness<br />

exercise described below, based on regional variation, is indicative<br />

and unavoidably rests on the validity of the identification<br />

assumption.<br />

We choose to exploit regional variability in the share of financial<br />

services over GDP, and merge our two data sets with administrative<br />

data available from the German Federal Statistical Office, which provides<br />

a broad set of structural data on some 500 local areas. The system<br />

of German zip codes is more granular than the regional grid<br />

used by the Federal Statistical Office. Accordingly, we map the zip<br />

codes for customer accounts into the regional grid of the Statistical<br />

Office by assuming that zip-codes from the same region share the<br />

same structural characteristics.<br />

We motivate the use of the GDP share of financial services by<br />

reference to the potential role that the density of financial services<br />

plays in reducing the cost of gathering financial information: the<br />

greater such density, the more likely it is that investors are able<br />

to gather information from local sources more cheaply, substituting<br />

for financial advice. Our identification assumption is that regional<br />

proximity with financial intermediaries affects account<br />

performance by facilitating the matching of account holders to<br />

financial advisors, but not directly. The instrument is collinear with<br />

the zip code dummies, which cannot be used in the estimation.<br />

Note however that in the IV estimates we still use wider regional<br />

dummies.<br />

Table A1 reports the first-stage estimates for IFA and for BFA,<br />

and shows that being located in an area with a larger density of<br />

intermediaries reduces the probability of having an account run<br />

by an IFA or a BFA. In both cases, the coefficients are statistically<br />

different from zero (at the 1% and 5% level, respectively).<br />

The IV estimates for portfolio performance are reported in Table<br />

A2. They are based on a standard IV estimator (with a linear probability<br />

model in the first stage) and standard errors adjusted for<br />

clustering at the local area level. With only one instrument, the<br />

model is exactly identified and we cannot provide a test of overidentification<br />

restrictions. However, we do find that the instrument<br />

has statistically significant impact on use of IFA (the F-statistic is<br />

reported in the last row of Table A2). The signs of the instrumented<br />

Table A1<br />

First-stage results: the determinants of having the account run by an IFA or a BFA.<br />

Brokerage sample<br />

(IFAs)<br />

Male 0.063 ***<br />

Bank sample<br />

(BFAs)<br />

0.049 ***<br />

(12.32) (3.16)<br />

Employee 0.058 **<br />

0.048 ***<br />

(2.49) (2.61)<br />

30 < Age 6 40 0.030 ***<br />

0.049<br />

(3.11) (1.41)<br />

40 < Age 6 50 0.002 0.011<br />

(0.25) (0.33)<br />

50 < Age 6 60 0.023 **<br />

0.055 *<br />

(2.21) (1.74)<br />

Age > 60 0.089 ***<br />

0.113 ***<br />

(7.62) (3.89)<br />

Log account volume 0.046 ***<br />

0.028 ***<br />

(22.60) (7.73)<br />

Self-employed 0.062 **<br />

(2.56)<br />

Experience/100 0.165 ***<br />

(3.58)<br />

Married 0.025 ***<br />

(6.06)<br />

Executive 0.001<br />

(0.01)<br />

Housewife 0.005<br />

(0.18)<br />

Retired 0.022<br />

(0.95)<br />

Financial services/regional<br />

GDP<br />

0.238 ***<br />

0.324 **<br />

(6.58) (2.53)<br />

Constant 0.301 ***<br />

0.446 ***<br />

(9.85) (7.54)<br />

Observations 28,321 4440<br />

R-squared 0.10 0.05<br />

The table reports estimates from a linear probability model of having an independent<br />

financial advisor or a bank financial advisor. Log account volume is measured<br />

in January 2003. The regressions include regional dummies. Asymptotic standard<br />

errors corrected for clustering at the zip code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.


Table A2<br />

Instrumental variable regressions for the brokerage and bank sample.<br />

Brokerage sample (IFAs) Bank sample (BFAs)<br />

Log monthly returns Variance of monthly returns Sharpe ratio Log monthly returns Variance of monthly returns Sharpe ratio<br />

(1) (2) (3) (4) (5) (6)<br />

Financial advisor 0.012 ***<br />

0.002 **<br />

0.166 ***<br />

-0.023 ***<br />

0.003 0.529 *<br />

(3.68) (2.54) (3.27) (2.60) (0.34) (1.73)<br />

Male 0.001 ***<br />

0.000 ***<br />

0.029 ***<br />

0.000 0.001 **<br />

0.048 **<br />

(4.20) (6.99) (7.64) (0.38) (2.27) (2.30)<br />

Employee 0.000 0.000 **<br />

0.015 0.000 0.001 0.026<br />

(0.54) (2.14) (1.31) (0.12) (1.17) (1.12)<br />

30 < Age 6 40 0.000 0.000 0.003 0.000 0.000 0.017<br />

(0.80) (1.08) (0.53) (0.43) (0.18) (0.49)<br />

40 < Age 6 50 0.000 0.000 ***<br />

0.013 ***<br />

0.000 0.002 *<br />

0.019<br />

(0.44) (5.31) (2.59) (0.06) (1.89) (0.64)<br />

50 < Age 6 60 0.000 0.000 ***<br />

0.011 **<br />

0.001 0.000 0.041<br />

(0.69) (5.97) (2.10) (0.59) (0.32) (1.16)<br />

Age > 60 0.000 0.001 ***<br />

0.009 0.000 0.001 0.067<br />

(0.44) (5.57) (1.21) (0.34) (0.37) (1.45)<br />

Log account volume 0.001 ***<br />

0.000 ***<br />

0.028 ***<br />

0.001 ***<br />

0.001 ***<br />

0.032 ***<br />

(4.66) (11.75) (11.60) (4.37) (4.09) (3.42)<br />

Self-employed 0.001 0.001 ***<br />

0.027 **<br />

(0.86) (5.23) (2.31)<br />

Experience/100 0.003 ***<br />

0.000 0.073 ***<br />

(2.97) (0.39) (4.24)<br />

Married 0.000 0.000 ***<br />

0.004<br />

(0.58) (7.58) (1.53)<br />

Executive 0.001 0.002 0.014<br />

(0.95) (1.47) (0.32)<br />

Housewife 0.001 0.000 0.009<br />

(1.13) (0.18) (0.34)<br />

Retired 0.001 0.000 0.019<br />

(0.90) (0.13) (0.83)<br />

Constant 0.005 ***<br />

0.007 ***<br />

0.027 0.009 **<br />

0.012 ***<br />

0.325 ***<br />

(4.02) (18.28) (1.20) (2.31) (3.28) (2.72)<br />

Observations 28,321 28,321 28,321 4440 4440 4440<br />

Cragg–Donald F-stat. 43.30 43.30 43.30 11.23 11.23 11.23<br />

The instrument is the GDP ratio of the financial asset share at the zip code level. The first stage regressions are reported in column 1 (brokerage sample) and column 2 (bank<br />

sample) of Table A1. Log account volume is measured in January 2003. All regressions include regional dummies. Asymptotic t-statistics corrected for clustering at the zip<br />

code level are reported in parentheses.<br />

* Statistical significance at the 10% level.<br />

** Statistical significance at the 5% level.<br />

*** Statistical significance at the 1% level.<br />

advisor effects on returns, variance and Sharpe ratios remain qualitatively<br />

unchanged as compared to the OLS results, for both the<br />

brokerage and bank sample. In particular, we find that advised accounts<br />

yield lower returns and Sharpe ratios, and that the negative<br />

effects are larger in the bank sample. Moreover, factors such as<br />

being female, experienced, and wealthier still contribute to higher<br />

returns, lower variance and higher Sharpe ratios.<br />

References<br />

Anderson, A., Martinez, J.V., 2008. Vice or advice? <strong>Prof</strong>its from brokerage firm<br />

trading around recommendation revision dates. Working paper, Stockholm<br />

University.<br />

Barber, B.M., Loeffler, D., 1993. The ‘Dartboard’ column: second-hand information<br />

and price pressure. Journal of Financial and Quantitative Analysis 28, 273–284.<br />

Barber, B.M., Odean, T., 2000. Trading is hazardous to your wealth: the common<br />

stock investment performance of individual investors. Journal of <strong>Finance</strong> 55,<br />

773–806.<br />

Barber, B.M., Odean, T., 2001. Boys will be boys: gender, overconfidence, and<br />

common stock investment. Quarterly Journal of Economics 116, 261–292.<br />

Barber, B.M., Lehavy, R., McNichols, M., Trueman, B., 2001. Can investors profit from<br />

the prophets? Security analysts recommendations and stock returns. Journal of<br />

<strong>Finance</strong> 56, 531–563.<br />

Bergstresser, D., Chalmers, J., Tufano, P., 2009. Assessing the costs and benefits of<br />

brokers: a preliminary analysis of the mutual fund industry. Review of Financial<br />

Studies 22, 4129–4156.<br />

Campbell, J.Y., 2006. Household finance. Journal of <strong>Finance</strong> 61, 1553–1604.<br />

Campbell, J.Y., Viceira, L.M., 2003. Strategic Asset Allocation – Portfolio Choice for<br />

Long-Term Investors. Oxford University Press, New York.<br />

A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524 523<br />

Carhart, M.M., 1997. On the persistence in mutual fund performance. Journal of<br />

<strong>Finance</strong> 52, 57–82.<br />

Cowles, A., 1933. Can stock market forecasters forecast? Econometrica 1, 309–324.<br />

Desai, H., Jain, P.C., 1995. An analysis of the recommendations of the ‘Superstar’ money<br />

managers at Barron’s annual roundtable. Journal of <strong>Finance</strong> 50, 1257–1274.<br />

Dietz, P.O., 1968. Components of a measurement model: rate of return, risk, and<br />

timing. Journal of <strong>Finance</strong> 23, 267–275.<br />

Dorn, D., Huberman, G., 2005. Talk and action: what individual investors say and<br />

what they do. Review of <strong>Finance</strong> 9, 437–481.<br />

Feng, L., Seasholes, M., 2005. Do investor sophistication and trading experience<br />

eliminate behavioral biases in financial markets? Review of <strong>Finance</strong> 9, 305–351.<br />

Goetzmann, W.N., Kumar, A., 2008. Equity portfolio diversification. Review of<br />

<strong>Finance</strong> 12, 433–463.<br />

Grinblatt, M., Keloharju, M., 2001. What makes investors trade? Journal of <strong>Finance</strong><br />

56, 589–616.<br />

Guiso, L., Jappelli, T., 2006. Information acquisition and portfolio performance. CSEF<br />

Working Paper No. 167.<br />

Inderst, R., Ottaviani, M., 2009. Misselling through agents. American Economic<br />

Review 99, 883–908.<br />

Lakonishok, J., Shleifer, A., Vishny, R.W., Hart, O., Perry, G.L., 1992. The structure and<br />

performance of the money management industry. Brookings Papers on<br />

Economic Activity: Microeconomics, 339–391.<br />

Lusardi, A., Mitchell, O.S., 2007. Baby boomer retirement security: the roles of<br />

planning, financial literacy, and housing wealth. Journal of Monetary Economics<br />

54, 205–224.<br />

Metrick, A., 1999. Performance evaluation with transactions data: the stock<br />

selection of investment newsletters. Journal of <strong>Finance</strong> 54, 1743–1774.<br />

Odean, T., 1998. Are investors reluctant to realize their losses? Journal of <strong>Finance</strong><br />

53, 1775–1798.<br />

Odean, T., 1999. Do investors trade too much? American Economic Review 89,<br />

1279–1298.<br />

Shapira, Z., Venezia, I., 2001. Patterns of behavior of professionally managed and<br />

independent investors. Journal of Banking and <strong>Finance</strong> 25, 1573–1587.


524 A. Hackethal et al. / Journal of Banking & <strong>Finance</strong> 36 (2012) 509–524<br />

Shu, P., Chiu, S., Chen, H., Yeh, Y., 2004. Does trading improve individual investor<br />

performance? Review of Quantitative <strong>Finance</strong> and Accounting 22, 199–217.<br />

Stulz, R.M., Mehran, H., 2007. The economics of conflicts of interest in financial<br />

institutions. Journal of Financial Economics 85, 267–296.<br />

Womack, K.L., 1996. Do brokerage analysts recommendations have investment<br />

value? Journal of <strong>Finance</strong> 51, 137–167.<br />

Venezia, I., Nashikkar, A., Shapira, Z., 2011. Firm specific and macro herding by<br />

professional and amateur investors and their effects on market volatility.<br />

Journal of Banking and <strong>Finance</strong> 35, 1599–1609.<br />

Yoong, J., Hung A., 2009. Self-dealing and compensation for financial advisors. RAND<br />

Working Paper Series, WR-713.


Philipp Schmitt, Bernd Skiera, & Christophe Van den Bulte<br />

© 2011, American Marketing Association<br />

ISSN: 0022-2429 (print), 1547-7185 (electronic)<br />

Referral Programs and<br />

Customer Value<br />

Referral programs have become a popular way to acquire customers. Yet there is no evidence to date that<br />

customers acquired through such programs are more valuable than other customers. The authors address this gap<br />

and investigate the extent to which referred customers are more profitable and more loyal. Tracking approximately<br />

10,000 customers of a leading German bank for almost three years, the authors find that referred customers (1)<br />

have a higher contribution margin, though this difference erodes over time; (2) have a higher retention rate, and<br />

this difference persists over time; and (3) are more valuable in both the short and the long run. The average value<br />

of a referred customer is at least 16% higher than that of a nonreferred customer with similar demographics and<br />

time of acquisition. However, the size of the value differential varies across customer segments; therefore, firms<br />

should use a selective approach for their referral programs.<br />

Keywords: customer referral programs, customer loyalty, customer value, customer management, word of mouth,<br />

social networks<br />

Word of mouth (WOM) has reemerged as an important<br />

marketing phenomenon, and its use as a customer<br />

acquisition method has begun to attract<br />

renewed interest (e.g., Godes and Mayzlin 2009; Iyengar,<br />

Van den Bulte, and Valente 2011). Traditionally, WOM’s<br />

appeal has been in the belief that it is cheaper than other<br />

acquisition methods. A few recent studies have documented<br />

that customers acquired through WOM also tend to churn<br />

less than customers acquired through traditional channels<br />

and that they tend to bring in additional customers through<br />

their own WOM (Choi 2009; Trusov, Bucklin, and Pauwels<br />

2009; Villanueva, Yoo, and Hanssens 2008). Villanueva,<br />

Yoo, and Hanssens (2008) further suggest that customers<br />

acquired through WOM can generate more revenue for the<br />

firm than customers acquired through traditional marketing<br />

efforts.<br />

From a managerial point of view, these findings are<br />

encouraging and suggest purposely stimulating WOM to<br />

acquire more customers. However, there are concerns that<br />

firm-stimulated WOM may be substantially less effective<br />

than organic WOM in generating valuable customers<br />

(Trusov, Bucklin, and Pauwels 2009; Van den Bulte 2010)<br />

because (1) targeted prospects may be suspicious of stimu-<br />

Philipp Schmitt is a doctoral student (e-mail: pschmitt@wiwi.uni-frankfurt.<br />

de), and Bernd Skiera is <strong>Prof</strong>essor of Marketing and Member of the Board<br />

of the E-<strong>Finance</strong> <strong>Lab</strong> at the House of <strong>Finance</strong> (e-mail: skiera@skiera.de),<br />

School of Business and Economics, Goethe University Frankfurt.<br />

Christophe Van den Bulte is Associate <strong>Prof</strong>essor of Marketing, the Wharton<br />

School, University of Pennsylvania (e-mail: vdbulte@wharton.upenn.<br />

edu). The authors thank the management of a company that wants to<br />

remain anonymous for making the data available and Christian Barrot,<br />

Jonah Berger, Xavier <strong>Dr</strong>èze, Peter Fader, Jeanette Heiligenthal, Gary<br />

Lilien, Renana Peres, Jochen Reiner, Christian Schulze, Russell Winer,<br />

Ezra Zuckerman, and the anonymous JM reviewers for providing comments<br />

on previous drafts of this article.<br />

46<br />

lated WOM efforts; (2) such efforts often involve a monetary<br />

reward for the referrer, who, as a result, may seem less<br />

trustworthy; (3) programs providing economic benefits tend<br />

not to be sustainable (Lewis 2006); (4) unlike organic<br />

WOM, stimulated WOM is not free, raising questions about<br />

cost effectiveness; and (5) stimulated WOM is prone to<br />

abuse by opportunistic referrers.<br />

The uncertainty about the benefits of stimulated WOM<br />

in customer acquisition is frustrating for managers facing<br />

demands to increase their marketing return on investment<br />

and considering whether to use this method. Our study<br />

addresses this managerial issue by investigating the value of<br />

customers acquired through stimulated WOM and comparing<br />

it with the value of customers acquired through other<br />

methods. We do so in the context of a specific WOM marketing<br />

practice that is gaining prominence, namely, referral<br />

programs in which the firm rewards existing customers for<br />

bringing in new customers. Although these programs are<br />

typically viewed as an attractive way to acquire customers,<br />

their benefits are often viewed to be their targetability and<br />

cost effectiveness (Mummert 2000). We broaden this view<br />

by assessing the value of customers acquired through these<br />

types of programs.<br />

Specifically, we answer four questions: (1) Are customers<br />

acquired through a referral program more valuable<br />

than other customers? (2) Is the difference in customer<br />

value large enough to cover the costs of such stimulated<br />

WOM customer acquisition efforts? (3) Are customers<br />

acquired through a referral program more valuable because<br />

they generate higher margins, exhibit higher retention, or<br />

both? and (4) Do differences in margins and retention<br />

remain stable, or do they erode? The answers to the last two<br />

questions provide deeper insight into what might be driving<br />

the value differential.<br />

We answer these four questions using panel data on all<br />

5181 customers that a leading German bank acquired<br />

Journal of Marketing<br />

Vol. 75 (January 2011), 46 –59


through its referral program (referred customers) between<br />

January 2006 and December 2006 and a random sample of<br />

4633 customers the same bank acquired through other<br />

methods (nonreferred customers) over the same period. For<br />

both groups of customers, we track profitability (measured<br />

as contribution margin) and loyalty (measured as retention)<br />

at the individual level from the date of acquisition until September<br />

2008. The total observation period spans 33 months.<br />

We use two metrics of customer value: (1) the present value<br />

of the actually observed contribution margins realized<br />

within the data window and (2) the expected present value<br />

over a period of six years from the day of acquisition.<br />

Although our study is limited to a single research site, as is<br />

common for studies that require rich and confidential data,<br />

the methodology and findings are of broad interest. Customer<br />

referral programs are gaining popularity in many<br />

industries, including financial services, hotels, automobiles,<br />

newspapers, and contact lenses (Ryu and Feick 2007).<br />

We make the following contributions: First, we provide<br />

empirical evidence that a referral program, a form of stimulated<br />

WOM, is an attractive way to acquire customers.<br />

Referred customers exhibit higher contribution margins,<br />

retention, and customer value. Second, building on our finding<br />

that differences in contribution margin erode over time<br />

whereas those in retention do not, we document that<br />

referred customers are more valuable in both the short and<br />

the long run. Third, we show that the referral effect need not<br />

be present in every customer segment. Finally, we illustrate<br />

how the type of analysis we conduct enables firms to calculate<br />

the return on investment and the upper bound for the<br />

reward in their customer referral programs.<br />

We proceed by describing referral programs and developing<br />

our hypotheses. A description of the research setting,<br />

the data, and the model specifications follows. Then, we<br />

report the results. Finally, we discuss implications for practice,<br />

the limitations, and opportunities for further research.<br />

Customer Referral Programs<br />

Customer referral programs are a form of stimulated WOM<br />

that provides incentives to existing customers to bring in<br />

new customers. An important requirement for such programs<br />

is that individual purchase or service histories are<br />

available so the firm can ascertain whether a referred customer<br />

is indeed a new rather than an existing or a former<br />

customer.<br />

Referral programs have three distinctive characteristics.<br />

First, they are deliberately initiated, actively managed, and<br />

continuously controlled by the firm, which is impossible or<br />

very difficult with organic WOM activities such as spontaneous<br />

customer conversations and blogs. Second, the key<br />

idea is to use the social connections of existing customers<br />

with noncustomers to convert the latter. Third, to make this<br />

conversion happen, the firm offers the existing customer a<br />

reward for bringing in new customers.<br />

Although leveraging the social ties of customers with<br />

noncustomers to acquire the latter is not unique to customer<br />

referral programs, the three distinctive characteristics of<br />

these programs set them apart from other forms of networkbased<br />

marketing (Van den Bulte and Wuyts 2007). Unlike<br />

organic WOM, the firm actively manages and monitors<br />

referral programs. Unlike most forms of buzz and viral marketing,<br />

the source of social influence is limited to existing<br />

customers rather than anyone who knows about the brand or<br />

event. Unlike multilevel marketing, existing customers are<br />

rewarded only for bringing in new customers. They do not<br />

perform any other sales function (e.g., hosting parties) and<br />

do not generate any income as a function of subsequent<br />

sales. Consequently, referral programs do not carry the<br />

stigma of exploiting social ties for mercantile purposes, as<br />

multilevel marketing does (Biggart 1989).<br />

In most referral programs, the reward is given regardless<br />

of how long the new referred customers stay with the firm.<br />

Such programs are prone to abuse by customers. Although<br />

the firm does not commit to accept every referral, the incentive<br />

structure combined with imperfect screening by the<br />

firm creates the potential for abuse in which existing customers<br />

are rewarded for referring low-quality customers.<br />

This kind of moral hazard is less likely to occur with WOM<br />

campaigns that do not involve monetary rewards conditional<br />

on customer recruitment.<br />

Existing studies of customer referral programs have<br />

provided guidance about when rewards should be offered<br />

(Biyalogorsky, Gerstner, and Libai 2001; Kornish and Li<br />

2010), have quantified the impact of rewards and tie<br />

strength on referral likelihood (Ryu and Feick 2007; Wirtz<br />

and Chew 2002), and have quantified the monetary value of<br />

making a referral (Helm 2003; Kumar, Petersen, and Leone<br />

2007, 2010). The key managerial question of the (differential)<br />

value of customers acquired through referral programs<br />

has not yet been addressed.<br />

Hypotheses<br />

Because referral programs are a customer acquisition<br />

method, an important metric to assess their effectiveness is<br />

the value of the customers they acquire. Additional insights<br />

come from investigating differences between referred and<br />

nonreferred customers in contribution margins and retention<br />

rates, the two main components of customer value (e.g.,<br />

Gupta and Zeithaml 2006; Wiesel, Skiera, and Villanueva<br />

2008).<br />

Our hypotheses regarding these customer metrics of<br />

managerial interest are informed by prior work in economics<br />

and sociology on employee referral (e.g., Coverdill 1998;<br />

Rees 1966), especially the work of Fernandez, Castilla, and<br />

Moore (2000), Neckerman and Fernandez (2003), and<br />

Castilla (2005) on the quality of employee referral programs.<br />

These studies show that the benefits of such programs are<br />

realized through distinct mechanisms, of which better<br />

matching and social enrichment appear particularly relevant<br />

to marketers. Better matching is the phenomenon that referrals<br />

fit with the firm better than nonreferrals, and social<br />

enrichment is the phenomenon that the relationship of the<br />

referral to the firm is enriched by the presence of a common<br />

third party (i.e., the referrer).<br />

Customer and employees referral programs are likely to<br />

be subject to similar mechanisms because they share the<br />

three distinctive characteristics of having active management,<br />

using the social connections of existing contacts, and<br />

Referral Programs and Customer Value / 47


offering rewards with the risk of abuse. Selecting a new<br />

employer or bank both are high-involvement decisions<br />

involving a fair amount of uncertainty. Although some basic<br />

banking products, such as checking accounts, are well<br />

known to most customers, the wider set of financial services<br />

that banks provide are considered experience goods rather<br />

than search goods (e.g., Bolton, Freixas, and Shapiro 2007;<br />

Parasuraman, Zeithaml, and Berry 1985). The recurrent<br />

losses of many private investors indicate that many people<br />

are not very skilled at assessing complex bank offerings.<br />

We use the better matching and social enrichment<br />

mechanisms to develop and motivate our hypotheses. However,<br />

our goal is to document managerially relevant differences<br />

in contribution margin, retention, and customer value<br />

rather than to test those specific mechanisms. The mechanisms<br />

are only possible explanations for the differences we<br />

document.<br />

Differences in Contribution Margin<br />

Several characteristics of social dynamics in human networks<br />

(e.g., Van den Bulte and Wuyts 2007) imply that<br />

referred customers may match up with the firm better than<br />

other newly acquired customers. The first is reciprocity.<br />

Because referring customers receive a reward, norms of reciprocity<br />

may make nonopportunistic customers feel obliged<br />

to bring in new customers who they think may be valuable<br />

to the firm (Gouldner 1960). This process explains Neckerman<br />

and Fernandez’s (2003) finding that referrals for which<br />

the referrer claimed a fee had lower turnover than referrals<br />

for which no fee was claimed. The second social dynamic<br />

likely to be at work is triadic balance. If the main function<br />

of the program is simply to nudge customers into making<br />

referrals without much consideration for the monetary<br />

reward (Thaler and Sunstein 2008), principles of triadic balance<br />

will make existing customers more likely to bring in<br />

others who they believe would match well with what the<br />

firm has to offer. A third social dynamic likely to be at work<br />

is homophily—the tendency for people to interact with<br />

people like them. Whereas reciprocity and triadic balance<br />

imply that referrers are diligent and active in screening and<br />

matching peers to firms, homophily implies that customers<br />

are likely to refer others who are similar to themselves.<br />

Because existing customers have an above-average chance<br />

of being a good match (otherwise, they would not be customers),<br />

firms may benefit from referral programs through<br />

“passive” homophily-based matching rather than only<br />

deliberate “active” screening-based matching by referrers<br />

(Kornish and Li 2010; Montgomery 1991; Rees 1966).<br />

Acquisition through referral can also result in informational<br />

advantages, making referred customers more profitable<br />

than other customers. Referred customers are likely<br />

to have discussed the firm’s offerings with their referrer. As<br />

a result, they are likely to use its products more extensively<br />

than novice customers who take a more cautious approach<br />

in building involvement. Informational advantages to the<br />

firm can also result if people refer others similar to themselves<br />

on dimensions that are relevant to the enjoyment of<br />

the product or service but are not immediately observable to<br />

the firm (Kornish and Li 2010). Examples for financial services<br />

include risk aversion and a sense of fiscal responsibil-<br />

48 / Journal of Marketing, January 2011<br />

ity. In these situations, the firm can make inferences from<br />

the observed behavior of the referrers about the products in<br />

which the referred customers will be most interested (e.g.,<br />

Guseva 2008). As a result, the firm is able to serve the<br />

referred customer in a tailored way early on, something that<br />

takes time to learn for other newly acquired customers.<br />

Because of this informational advantage, the firm should be<br />

able to generate a higher contribution margin from referred<br />

customers at the beginning of the relationship.<br />

However, the advantages of better matching and better<br />

information should gradually vanish. As nonreferred customers<br />

accumulate experience with the firm, they should<br />

become equally well informed about the firm’s offerings<br />

and procedures. Likewise, the firm should be able to use the<br />

purchase and service history of the nonreferred customers<br />

to serve them better. Furthermore, nonreferred customers<br />

who are not a good match for the firm are more likely to<br />

leave. Consequently, both revenues and costs of referred<br />

and nonreferred customer should converge, eliminating the<br />

difference in contribution margin over time. Thus:<br />

H 1: (a) The average contribution margin of a customer acquired<br />

through a referral program is higher than that of a customer<br />

acquired through other methods, but (b) this difference<br />

becomes smaller over time.<br />

Differences in Retention<br />

Social enrichment is another mechanism that may increase<br />

the value of referred customers. The argument is that the<br />

relationship with the firm is enriched because a family<br />

member or friend is a customer of the same firm (Castilla<br />

2005; Fernandez, Castilla, and Moore 2000). Having a person<br />

close to oneself in a similar position (i.e., being a customer<br />

of the same firm) should increase the person’s trust in<br />

the firm and strengthen his or her emotional bond with it, as<br />

both balance theory and social closure theory predict (Van<br />

den Bulte and Wuyts 2007). This prediction is also consistent<br />

with findings that customers reflecting on their affect toward<br />

a firm mention friends who are customers with the same<br />

firm (Yim, Tse, and Chan 2008). Such relationships should<br />

be particularly relevant in a banking context, in which emotions<br />

and trust play important roles in the customer–firm<br />

relationship (e.g., Edwards and Day 2005; Fleming, Coffman,<br />

and Harter 2005). In short, referred customers are<br />

likely to have a stronger sense of commitment and attachment<br />

to the firm. This implies that referred customers are<br />

less likely to churn than nonreferred customers, provided<br />

that their referrer does not churn either. The latter condition<br />

is likely to hold: Referrers typically have a greater longterm<br />

likelihood of staying, which is why intention to refer is<br />

frequently used as an indicator of loyalty (Gupta and Zeithaml<br />

2006).<br />

Although the informational advantage of a referred customer<br />

decreases over time as direct experience substitutes<br />

for social learning, there is no reason to expect erosion in<br />

the affective bonding underlying the social enrichment<br />

mechanism. Consequently, the erosion of the differential<br />

expected in contribution margin need not apply to retention.<br />

Therefore, we state the following:


H 2: (a) The average retention of a customer acquired through a<br />

referral program is higher than that of a customer acquired<br />

through other methods, and (b) this difference does not<br />

become smaller over time.<br />

Differences in Customer Value<br />

If H1 and H2 hold and if the erosion of contribution margins<br />

does not outweigh the initial difference in margins and the<br />

persisting difference in retention, the following should hold<br />

as well:<br />

H 3: The average value of a customer acquired through a referral<br />

program is higher than that of a customer acquired<br />

through other methods.<br />

H 3 can hold even when H 1 and H 2 do not, because it is possible<br />

for the differences in both margins and retention to be<br />

small but for their combined effect to be sizable and significant.<br />

Another reason to test H 3 on customer value, in addition<br />

to H 1 and H 2 on margins and retention, is that customer<br />

value is what managers should care about most.<br />

Although we base our prediction on sound theoretical<br />

arguments, the posited effects are not as obvious as they<br />

may seem because there are several competing forces at<br />

work. First, the prospect of earning a referral fee may<br />

induce referrers to pressure their peers to become customers<br />

without giving much consideration to whether the firm<br />

actually matches their peers’ needs. Second, the relationship<br />

between the referred customer and his or her referrer, which<br />

is necessary for social enrichment to operate, comes with an<br />

inherent risk: When referrers defect, the customers they<br />

brought in may become more likely to leave as well.<br />

Although it seems unlikely that referrers as a whole are<br />

more prone to churn than the average customer, the risk of<br />

contagious defection should not be ignored altogether.<br />

Third, an abuse of the referral program by customers who<br />

are interested only in the monetary reward is probably the<br />

most important reason for practitioners’ skepticism. This is<br />

illustrated by TiVo’s termination of its referral program as a<br />

result of frequent abuses (ZatzNotFunny 2008).<br />

Support for our hypotheses would allow us to conclude<br />

that the positive effects outweigh the negative ones. In addition,<br />

the empirical analysis provides not only a test of the<br />

hypotheses but also an estimate of the size of the customer<br />

value differential. Firms can use the latter to put a maximum<br />

value on the reward to be paid out as part of their<br />

referral program.<br />

Methods<br />

Research Setting<br />

We use data from a leading German bank, whose name we<br />

do not divulge for confidentiality reasons. The data capture<br />

all customers acquired through the bank’s referral program<br />

between January 2006 and December 2006 and a random<br />

sample of customers acquired through other methods (e.g.,<br />

direct mail, advertising) over the same period. The latter<br />

group may include customers affected by organic WOM for<br />

which the bank did not pay any fee. To the extent that the<br />

value of customers acquired through organic WOM is equal<br />

to or greater than that of customers acquired through the<br />

referral program, our results underestimate the value differential<br />

between WOM and non-WOM customers. Regardless,<br />

we correctly estimate the value differential between<br />

customers acquired through the referral program and all<br />

other customers for whom no referral fee was paid.<br />

The observation period spans from January 2006 to<br />

September 2008 (33 months), and the data on each individual<br />

customer include the day of acquisition, the day of leaving<br />

the bank (if applicable), the contribution margin in each<br />

year, and some demographics. In total, we have data on<br />

5181 referred and 4633 nonreferred customers. Because the<br />

referral program was used only in a business-to-consumer<br />

context, all customers are individual people.<br />

The referral program was communicated to existing<br />

customers through staff and flyers in the branches and<br />

through mailings. 1 The procedure was straightforward:<br />

Every existing customer who brought in a new customer<br />

received a reward of 25 euros in the form of a voucher that<br />

could be used at several well-known German retailers. 2<br />

Except for opening an account, the referred customer did not<br />

need to meet any prerequisites (e.g., minimum amount of<br />

assets, minimum stay) for the referrer to receive the reward.<br />

In addition, 2006 was not an unusual year in terms of<br />

customer acquisition, and the bank’s management was confident<br />

that findings about customers acquired in 2006 would<br />

be applicable to customers acquired in earlier and later<br />

years. Proprietary information of the bank shows that its<br />

customers are similar to those of other leading European<br />

banks. Regarding the usage of its referral program and the<br />

response of its customers to it, no differences with other<br />

firms are apparent either.<br />

Dependent Variables<br />

Daily contribution margin is the individual contribution<br />

margin on a daily basis. It is the total contribution margin<br />

the customer generates in the observation period, divided by<br />

the total number of days the customer was with the bank<br />

(duration). This per diem scaling ensures the comparability<br />

of the contribution margin of customers with different<br />

observed (and possible censored) durations. The contribution<br />

margin equals revenue (interest and fees) less direct<br />

costs (e.g., interest expenses, sales commissions, brokerage,<br />

trading costs). The acquisition costs are not subtracted from<br />

the contribution margin. We also compute a time-varying<br />

version of daily contribution margin by dividing the contribution<br />

margin generated by the customer in a particular<br />

year (2006, 2007, 2008) by the number of days the customer<br />

was with the bank in that year.<br />

1These mailings went to the referring customers. Mailings to<br />

which the nonreferred customers responded were sent directly to<br />

them.<br />

2Although confidentiality concerns preclude us from reporting<br />

the average cost of acquisition for referral and nonreferral customers,<br />

we can report that the total acquisition cost for referred<br />

customers (including not only the referral fee but also the additional<br />

administrative costs of record keeping, paying out, and so<br />

on) is on average approximately 20 euros lower than that for nonreferred<br />

customers acquired through mailings.<br />

Referral Programs and Customer Value / 49


Duration is the total number of days the customer was<br />

observed to be with the bank. It is the basis for analyzing<br />

retention.<br />

We calculate two measures of customer value, one<br />

based only on observed data and the other based on both<br />

observed data and predictions. Observed customer value is<br />

the present value of all actual contribution margins the customer<br />

generated during the whole observation period (e.g.,<br />

33 months for retained customers acquired in January<br />

2006). This metric is affected by both contribution margin<br />

and retention because a customer generates no margins after<br />

leaving the bank. Our second metric, customer lifetime<br />

value, is the present value of all contribution margins, both<br />

actual and predicted, of the customer within the six-year<br />

span following the day of acquisition. 3 For customers who<br />

churned during the observation period, customer lifetime<br />

value equals observed customer value because their predicted<br />

value is 0. The formulas are as follows:<br />

(1) Customer Lifetime Value i = Observed Customer Value i<br />

( 2)<br />

Observed Customer Valuei =<br />

+ Predicted Customer Value i,<br />

Duri<br />

∑<br />

s = 1<br />

( 3)<br />

Predicted Customer Valuei = δi<br />

OM<br />

( 1 + r)<br />

is<br />

s 12<br />

, and<br />

where OM is is the observed monthly contribution margin of<br />

customer i in the sth month after acquisition (calculated<br />

from the observed annual contribution margin and the<br />

observed duration), Dur i is the customer’s observed lifetime<br />

with the bank in months, d i is a dummy censoring variable<br />

that indicates whether the customer was still with the bank<br />

by the end of the observation period, PM is is the predicted<br />

monthly contribution margin of customer i in the sth month<br />

after acquisition, PA is is the predicted probability that customer<br />

i is still “alive” (i.e., with the bank) in that month,<br />

and r is the firm-specific annual discount rate of 11.5%. 4<br />

The present value reflects what the customer is worth at the<br />

day of acquisition.<br />

Independent Variables<br />

The independent variable of central interest is referral program,<br />

a binary indicator that takes the value 1 for referred<br />

customers (i.e., those who were acquired through the referral<br />

program) and 0 for nonreferred customers. To improve the<br />

comparability of referred and nonreferred customers, we control<br />

for the demographic variables age, sex (dummy variable,<br />

s<br />

72<br />

∑<br />

= Duri + 1<br />

50 / Journal of Marketing, January 2011<br />

PMis × PAis,<br />

( 1 + r)<br />

s 12<br />

3This way, we do not need to predict margins and retention rates<br />

beyond four years after the end of the data window, and the resulting<br />

customer lifetime values are unlikely to be overly affected by<br />

forecasting error (Kumar and Shah 2009).<br />

4We base the discount rate on the capital asset pricing model.<br />

We assume a risk-free interest rate of 4.25% (Svensson 1994), a<br />

5% market risk rate based on the Institute of German Accountants,<br />

and a firm-specific beta of 1.45 based on Thomson Financial<br />

Datastream.<br />

with women coded as 1 and men coded as 0), and marital<br />

status (dummy variables for married, divorced/separated,<br />

single, and widowed, with no answer as the base category).<br />

We also control for the customer’s month of acquisition (11<br />

dummy variables for each month, with December 2006 as<br />

the base category).<br />

To assess the robustness of the difference in customer<br />

value, we also conduct separate analyses for the two key<br />

segments of the bank: retail customers with standard financial<br />

needs and nonretail customers with significant assets or<br />

requiring more sophisticated financial advice. This segmentation<br />

scheme the bank uses is based on a comprehensive<br />

analysis of both financial data (e.g., assets invested with the<br />

bank, monthly checking account balance) and demographic<br />

information (e.g., profession, place of residence). The segments<br />

form the basis for all strategic customer-related decisions<br />

of the bank.<br />

Descriptive Statistics<br />

The sample includes several customers with an extremely<br />

high daily contribution margin that is up to 80 standard<br />

deviations above the mean and median. Such extreme data<br />

points can influence comparisons of means and regression<br />

results, so we purify the data using the standard DFBETA<br />

and DFFIT criteria to identify influence points (Belsley,<br />

Kuh, and Welsch 1980). This procedure led to the deletion<br />

of 172 referred customers (3.3% of the original 5181<br />

referred customers) and 147 nonreferred customers (3.2%<br />

of the original 4633 nonreferred customers). As we report in<br />

the subsection “Robustness Checks,” testing the hypotheses<br />

without deleting the influence points results in larger differences<br />

and provides stronger support for the hypotheses. Yet<br />

the size estimates obtained without the influence points are<br />

more informative.<br />

Table 1 presents the means, standard deviations, and the<br />

correlation matrix for the purified sample of 9495 customers.<br />

As the nonzero correlations between the referral<br />

program variable and the various demographic and time of<br />

acquisition variables indicate, the groups of referred and<br />

nonreferred customers are not perfectly matched on demographics<br />

and time of acquisition. Thus, it is desirable to<br />

control for these differences.<br />

Figure 1 plots the average daily contribution margin for<br />

the referred and nonreferred customers of the purified sample,<br />

for 2006, 2007, and 2008. The pattern is encouraging.<br />

Referred customers tend to generate higher margins, and<br />

the margins tend to erode more quickly among referred customers,<br />

such that the margin differential is narrower in 2008<br />

than in 2006 (8 cents versus 18 cents per day). Similarly, as<br />

Figure 2 shows, after about a year, the retention rate of<br />

referred customers is higher, and this is the case regardless<br />

of duration. However, controlling for differences in demographics<br />

and time of acquisition is necessary to draw conclusions<br />

more confidently.<br />

Statistical Models<br />

To estimate the difference in contribution margin between<br />

acquisition modes (H1a), we use a regression model with<br />

the following specification:


TAble 1<br />

Descriptive Statistics<br />

Units M SD 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22<br />

1. Referral<br />

program 0–1 .53 .50 1.00<br />

2. Observed<br />

customer<br />

value Euros 210.66 336.15 .02 1.00<br />

3. Customer<br />

lifetime value Euros 255.75 338.95 .01 1.00 1.00<br />

4. Daily<br />

contribution Euros<br />

margin per day .33 .50 .04 .98 .98 1.00<br />

5. Duration Days 751.05 119.48 –.17 .18 .21 .09 1.00<br />

6. Age Years 42.90 17.47 –.20 .10 .09 .10 –.01 1.00<br />

7. Female 0–1 .54 .50 .07 .01 .01 .01 .01 .05 1.00<br />

8. Married 0–1 .39 .49 –.15 –.02 –.03 –.02 –.03 .43 .01 1.00<br />

9. Single 0–1 .44 .50 .16 –.05 –.04 –.05 .01 –.65 –.10 –.70 1.00<br />

10. Divorced 0–1 .10 .30 .00 .03 .03 .03 .02 .13 .06 –.26 –.29 1.00<br />

11. Widowed 0–1 .05 .22 –.05 .11 .11 .10 .03 .36 .14 –.18 –.20 –.07 1.00<br />

12. Acquired<br />

January 2006 0–1 .03 .17 –.17 .07 .08 .03 .31 .02 .00 .01 –.03 –.00 .04 1.00<br />

13. Acquired<br />

February 2006 0–1 .03 .18 –.18 .02 .03 –.01 .27 .05 –.02 .04 –.04 –.00 .01 –.03 1.00<br />

14. Acquired<br />

March 2006 0–1 .06 .24 –.18 .04 .04 .01 .29 .07 –.01 .05 –.04 –.00 .00 –.04 –.05 1.00<br />

15. Acquired<br />

April 2006 0–1 .06 .23 .02 .04 .04 .02 .24 –.01 –.00 –.03 .02 .02 .01 –.04 –.05 –.06 1.00<br />

16. Acquired<br />

May 2006 0–1 .07 .26 .03 .03 .04 .01 .22 –.02 –.01 –.01 .02 –.01 –.02 –.05 –.05 –.07 –.07 1.00<br />

17. Acquired<br />

June 2006 0–1 .08 .28 –.01 .02 .02 .01 .14 .02 .02 –.01 –.01 .00 .04 –.05 –.06 –.08 –.07 –.08 1.00<br />

18. Acquired<br />

July 2006 0–1 .10 .30 .00 .01 .01 .01 .08 .02 –.00 .00 –.02 .03 –.00 –.06 –.06 –.09 –.08 –.09 –.10 1.00<br />

19. Acquired<br />

August 2006 0–1 .11 .31 .06 –.00 –.00 –.00 –.01 –.08 .00 –.06 .05 .02 –.02 –.06 –.06 –.09 –.08 –.09 –.10 –.12 1.00<br />

20. Acquired<br />

September 2006 0–1 .08 .27 .07 .00 .00 .01 –.08 –.06 .01 –.03 .04 –.00 –.02 –.05 –.06 –.08 –.07 –.08 –.09 –.10 –.10 1.00<br />

21. Acquired<br />

October 2006 0–1 .12 .33 .04 –.03 –.04 –.01 –.22 .01 .01 .02 –.00 –.01 –.01 –.07 –.07 –.09 –.09 –.10 –.11 –.13 –.13 –.11 1.00<br />

22. Acquired<br />

November 2006 0–1 .12 .33 .04 –.05 –.06 –.02 –.31 –.00 –.03 .01 –.01 –.01 –.00 –.07 –.07 –.09 –.09 –.10 –.11 –.13 –.13 –.11 –.14 1.00<br />

23. Nonretail<br />

customers 0–1 .12 .32 –.03 .27 .27 .27 –.00 .03 .00 .01 –.01 –.03 .00 .02 –.00 .01 .01 –.04 .00 –.01 –.02 –.02 –.03 .00<br />

Notes: N = 9495. All correlations with absolute value of .02 or higher are significant at the 5% level. Differences in observed duration across customers are strongly affected by differences in the<br />

month of acquisition. As a result, the zero-order correlations of duration with other variables that are also correlated with month of acquisition can be misleading. For example, the correlation<br />

between duration and referral program changes from –.17 to .03 after we control for month of acquisition.<br />

Referral Programs and Customer Value / 51


Daily Contribution Margin (€)<br />

Probability<br />

FIGURe 1<br />

Average Values of Daily Contribution Margin for<br />

Referred and Nonreferred Customers by Year<br />

.600<br />

.500<br />

.400<br />

.300<br />

.200<br />

.100<br />

.496<br />

.315<br />

.331<br />

.350<br />

∑<br />

( ) DCM = α + β RP + β X + ε ,<br />

4 i 1 i k<br />

k = 2<br />

ik i<br />

where i indexes the customer, DCM is daily contribution<br />

margin over the observation period, RP is the binary variable<br />

representing the referral program, X represents the control<br />

variables, and the errors e i have a mean of zero and are<br />

independent of the included covariates. We use ordinary least<br />

squares to estimate the coefficients and compute Huber–<br />

White heteroskedasticity-consistent standard errors (Breusch–<br />

Pagan test, p < .001). The size of our sample implies that<br />

we do not need to assume that the random shocks are normally<br />

distributed for statistical inference using t- and F-statistics<br />

(e.g., Wooldridge 2002, pp. 167–71).<br />

To test whether difference in margin decreases the longer<br />

the customer has been with the bank (H 1b), we use a fixedeffects<br />

specification estimated with ordinary least squares:<br />

( ) DCM = α + β T + β RP × T + η + ε ,<br />

5 it i 2 it 3 i it t it<br />

52 / Journal of Marketing, January 2011<br />

Referred<br />

Nonreferred<br />

.269<br />

.189<br />

2006 2007<br />

Year<br />

2008<br />

FIGURe 2<br />

Probability That Referred and Nonreferred<br />

Customers Have Remained with the Firm<br />

(Kaplan–Meier estimates of Survivor Functions)<br />

1.00<br />

.95<br />

.90<br />

.85<br />

.80<br />

.75<br />

Referred Nonreferred<br />

0 200 400 600 800 1000<br />

Days Since Acquisition<br />

Notes: Customers were able to leave immediately after joining, but<br />

only a handful did so. The earliest defection took place after<br />

64 days, and only 27 customers left within the first year of<br />

joining.<br />

where i indexes the customer; t indexes the year (t = 1, 2, 3);<br />

DCM it is the daily contribution margin of customer i in year<br />

t (i.e., the total contribution margin generated by customer i<br />

in year t divided by the number of days the customer was<br />

with the firm during year t); T it is the cumulative number of<br />

days customer i had been with the bank by the end of year t;<br />

h t is a year-specific fixed effect; and the customer-specific<br />

intercepts a i are not constrained to follow any specific distribution<br />

but capture all individual-specific, time-invariant<br />

differences, including the effect of acquisition through the<br />

referral program (RP) and that of the control variables X.<br />

The errors e it have a mean of zero and are independent of<br />

the covariates. The b 3 coefficient captures the proper interaction<br />

effect because the b 1 effect of RP is now captured<br />

through the fixed effects. Again, we use the robust<br />

Huber–White standard errors (Breusch–Pagan test, p < .001).<br />

To assess the difference in retention between acquisition<br />

modes, we use the Cox proportional hazard model. This<br />

enables us to analyze right-censored duration data and to<br />

exploit the fine-grained measurement of churn at the daily<br />

level without imposing any restriction on how the average<br />

churn rate evolves over time. Furthermore, the nonparametric<br />

baseline hazard makes the model robust to unobserved<br />

heterogeneity in all but extreme cases (e.g., Meyer 1990).<br />

We can represent the model to test H 2a as follows:<br />

∑<br />

( ) ln[ h ( t)] = α( t) + β RP + β X ,<br />

6 i 1 i k ik<br />

k = 2<br />

where i indexes the customer, t indexes the amount of days<br />

since the customer joined the bank, h i(t) is the hazard rate<br />

for the customer’s defection, and a(t) is the log of the nonparametric<br />

baseline hazard common across all customers.<br />

To test whether the difference in churn propensity changes<br />

over time (H 2b), we extend Equation 6 with the interaction<br />

between RP i and ln(t). The latter is also a test of whether<br />

the RP effect meets the proportionality assumption (e.g.,<br />

Blossfeld, Hamerle, and Mayer 1989), but we use it here to<br />

test a hypothesis of substantive interest.<br />

To test H 3 and assess the difference in customer value,<br />

we again use the regression model in Equation 4, but now<br />

with observed customer value as the dependent variable.<br />

Theoretical claims can be subjected to empirical validation<br />

or refutation only by comparing hypothesized effects with<br />

actual data. As a result, the truth content of H 3 can be validly<br />

tested using the observed customer value as the dependent<br />

variable but not customer lifetime value, which itself is based<br />

on predictions. Still, given the right censoring of our data<br />

and the hypothesized erosion of the margin differential over<br />

time, it is informative also to perform a similar analysis with<br />

the six-year customer lifetime value as the dependent variable.<br />

To calculate the predicted values entering the customer lifetime<br />

value metric, we use (1) predicted annual contribution<br />

margins from a fixed-effects model, such as that specified<br />

in Equation 5, but one in which we allow all parameters to<br />

vary between referred and nonreferred customers, and (2)<br />

predicted annual survival rates from a parametric Weibull<br />

hazard model because the nonparametric baseline hazard of<br />

the Cox model does not allow for forecasts. 5


Results<br />

Is the Contribution Margin of Referred Customers<br />

Higher?<br />

In accordance with H1a, referred customers are, on average,<br />

4.5 cents per day more profitable than other customers<br />

(Mann–Whitney test, p < .001). The difference is even larger<br />

after we control for differences in customer demographics<br />

and time of acquisition, variables on which the two groups<br />

of customers are not perfectly matched. The first column of<br />

Table 2 reports the coefficients of Equation 4, controlling<br />

for age, sex, marital status, and month of acquisition.<br />

Whereas the average contribution margin of nonreferred<br />

customers in our sample is 30 cents per day, customers<br />

acquired through the referral program have a margin that is<br />

7.6 cents per day or 27.74 euros per year higher (p < .001),<br />

an increase of approximately 25%. Among the covariates,<br />

older age and being widowed are associated with a higher<br />

contribution margin, whereas being married is associated<br />

with a lower contribution margin. The pattern in the<br />

monthly coefficients suggests that the bank was more successful<br />

in acquiring profitable customers in some months<br />

than in others. The R-square is low, indicating that important<br />

elements other than acquisition method, acquisition<br />

time, and demographics drive customer profitability.<br />

Although the large unexplained variance depresses the<br />

power of statistical tests and thus makes it more difficult to<br />

reject the null hypothesis, H1a is strongly supported.<br />

Does the Contribution Margin of Referred<br />

Customers Remain Higher?<br />

H1b predicts that the difference in contribution margin<br />

erodes the longer a customer stays with the bank. The<br />

results support this. Column 2 of Table 2 reports the coefficients<br />

of the fixed-effects model in Equation 5. There is a<br />

significant, negative interaction between referral program<br />

and the number of days the customer has been with the<br />

bank. The difference in daily contribution margin between<br />

referred and nonreferred customers decreases by 23.1 cents<br />

per 1000 days, or 8.4 cents per year.<br />

The individual-level fixed effects (intercepts) in the<br />

model capture the expected daily contribution margin when<br />

the included covariates equal zero (i.e., on the day of acquisition<br />

in 2006). Regression of these 9495 fixed-effects estimates<br />

on the referral program and control variables indicates<br />

that a referred customer has an expected contribution<br />

margin on the first day of joining the firm that is 19.8 cents<br />

higher than that of a nonreferred customer with similar<br />

demographics and time of acquisition. 6 This implies that<br />

5In-sample parameter estimates from the Cox and Weibull models<br />

are nearly identical. The reason for using the Cox model in<br />

testing the hypotheses is the absence of a restrictive parametric<br />

assumption on the duration dependence.<br />

6This difference in daily contribution margin of 19.8 cents is<br />

close but not identical to the 18 cents difference between the two<br />

groups of customers in 2006, shown in Figure 1. The small disparity<br />

between the two values occurs because the former controls for<br />

differences in demographics and time of acquisition, whereas the<br />

latter does not. Another reason for the disparity is that the former<br />

is the difference on the day of acquisition, whereas the latter is the<br />

difference on an average day in 2006.<br />

the expected contribution margin advantage of a referred<br />

customer disappears after 857 days (.198/.000231), or<br />

approximately 29 months after the customer joined the<br />

bank.<br />

Is the Retention of Referred Customers Higher?<br />

To test whether the retention rate is higher for referred than<br />

for nonreferred customers (H2a), we use the Cox proportional<br />

hazard model specified in Equation 6. The results<br />

reveal that the association between referral program and<br />

churn (i.e., nonretention) is significantly negative and sizable.<br />

Use of only referral program as the explanatory<br />

variable shows that at any point in time, customers acquired<br />

through the referral program who are still with the firm are<br />

approximately 13% less likely to defect than nonreferred<br />

customers who are still with the firm. After we control for<br />

differences in demographics and month of acquisition (see<br />

Column 3 of Table 2), the effect size increases to approximately<br />

–18% [exp(–.198) – 1]. This multiplicative effect of<br />

18% is relative to a baseline hazard that is small. As the survival<br />

curves in Figure 2 show, the probability of being an<br />

active customer (i.e., a nonchurning customer) after 33<br />

months is 82.0% for referred customers and 79.2% for nonreferred<br />

customers. Age is associated with a higher churn<br />

rate, whereas the opposite holds for being widowed. There<br />

is also a trend in the monthly coefficients, indicating that<br />

customers acquired late in 2006 (especially in September<br />

and later) exhibit more churn than those acquired earlier.<br />

This trend is a cohort effect and not duration dependence,<br />

which is captured in the nonparametric baseline hazard.<br />

Does the Retention of Referred Customers<br />

Remain Higher?<br />

We also assess whether the difference in retention varies<br />

over the customer’s lifetime (H2b). To do so, we extend the<br />

Cox model with an interaction between the referral program<br />

variable and the natural logarithm of the customer’s duration<br />

with the bank (see Column 4 of Table 2). The interaction<br />

is not significant, and the model fit does not improve<br />

significantly (p > .05). So, although there is an eroding difference<br />

between referred and nonreferred customers in contribution<br />

margin, there is no such erosion for customer<br />

retention. 7<br />

Are Referred Customers More Valuable?<br />

Using the observed customer value, we find that, on average,<br />

referred customers are 18 euros more valuable<br />

(Mann–Whitney test, p < .001). After we control for demographics<br />

and month of acquisition, the difference increases<br />

to 49 euros (Column 5 of Table 2; p < .001). A referred cus-<br />

7Note that in the model with the interaction term included, the<br />

coefficient of referral program (.917, p > .05) is not the average<br />

difference between referred and nonreferred customers anymore<br />

but rather the difference between the two groups on the day of<br />

acquisition (i.e., when the interaction variable Log(duration)<br />

equals 0; see Irwin and McClelland 2001). Thus, the insignificant<br />

coefficient of referral program in the model including the interaction<br />

term does not invalidate the finding of a significant difference<br />

in retention between the two groups posited in H 2a.<br />

Referral Programs and Customer Value / 53


TAble 2<br />

Main Results for Differences in Daily Contribution Margin, Churn (i.e., the Converse of Retention),<br />

Observed Customer Value, and Customer lifetime Value<br />

H1b H1a Daily<br />

H2b H3 H3 Daily<br />

Contribution<br />

Contribution<br />

Margin (Time<br />

H2a Churn<br />

(Time<br />

Observed<br />

Customer<br />

Customer<br />

lifetime<br />

Margin Varying) Churn Varying) Value Value<br />

Referral program .076*** — a –.198** .917 49.157*** 39.906***<br />

(.010) — (.059) (1.479) (7.096) 7.152<br />

Age .003*** — .011** .011*** 1.879*** 1.626***<br />

(.000) — (.002) (.002) (.283) (.285)<br />

Female –.009 — –.034 –.034 –4.459 –3.376<br />

(.010) — (.056) (.056) (6.902) (6.958)<br />

Sarried –.078* — –.027 –.028 –52.798* –52.258*<br />

(.033) — (.166) (.166) (22.427) (22.563)<br />

Single –.040 — –.163 –.163 –27.306 –24.035<br />

(.033) — (.167) (.167) (22.573) (22.706)<br />

Divorced –.016 — –.176 –.177 –12.278 –7.656<br />

(.037) — (.183) (.183) (24.776) (24.933)<br />

Widowed .111* — –.470* –.470* 76.085* 87.249**<br />

(.046) — (.212) (.212) (31.128) (31.355)<br />

Acquired January 2006 .172*** — –1.828** –1.833*** 228.228*** 247.960***<br />

(.039) — (.201) (.201) (31.589) (31.666)<br />

Acquired February 2006 .063* — –1.365** –1.369*** 127.706*** 133.591***<br />

(.031) — (.160) (.159) (24.172) (24.411)<br />

Acquired March 2006 .089** — –1.155** –1.157*** 136.393*** 135.755***<br />

(.026) — (.126) (.126) (19.103) (19.280)<br />

Acquired April 2006 .084** — –1.215** –1.208*** 124.793*** 123.153***<br />

(.027) — (.140) (.140) (18.753) (18.895)<br />

Acquired May 2006 .082** — –1.529** –1.524*** 114.302*** 119.426***<br />

(.025) — (.150) (.150) (16.791) (16.909)<br />

Acquired June 2006 .066** — –1.016** –1.013*** 91.090*** 92.643***<br />

(.022) — (.122) (.122) (14.326) (14.475)<br />

Acquired July 2006 .062** — –1.026** –1.023*** 79.574*** 84.200***<br />

(.021) — (.122) (.122) (12.717) (12.839)<br />

Acquired August 2006 .059** — –.841** –.838*** 69.213*** 73.167***<br />

(.020) — (.119) (.119) (12.111) (12.233)<br />

Acquired September 2006 .077** — –.679** –.676*** 72.213*** 76.352***<br />

(.022) — (.126) (.126) (13.199) (13.335)<br />

Acquired October 2006 .037 — –.434** –.432*** 36.602* 39.391***<br />

(.020) — (.108) (.108) (11.133) (11.257)<br />

Acquired November 2006 .021 — –.217* –.215* 19.252 20.551<br />

(.019) — (.105) (.105) (10.497) (10.632)<br />

Year 2007 (dummy) –1.306<br />

(.732)<br />

Year 2008 (dummy) –2.259<br />

(1.258)<br />

Cumulative Days (in thousands)<br />

Cumulative days (in thousands) ¥<br />

3.513<br />

(1.994)<br />

referral program –.231**<br />

(.085)<br />

Log(duration) ¥ referral program –.176<br />

(.232)<br />

Constant .154*** 66.250* 120.949***<br />

(.040) (26.742) (26.937)<br />

Observations 9495 28,353 9495 9495 9495 9495<br />

R² .025 .350 .040 .040<br />

Log-pseudo-likelihood<br />

*p < .05.<br />

**p < .01.<br />

***p < .001.<br />

aCaptured by customer-specific fixed effects.<br />

Notes: Robust standard errors are in parentheses.<br />

–11,715.6 –11,715.4<br />

54 / Journal of Marketing, January 2011


tomer is approximately 25% more valuable to the bank than<br />

a comparable nonreferred customer, within the observation<br />

period. If we take into account the difference in acquisition<br />

costs of approximately 20 euros, the difference in customer<br />

value is nearly 35%. These results strongly support H 3.<br />

Because the margin differential of referred customers<br />

erodes over time even though the loyalty differential does<br />

not, the question arises whether referred customers remain<br />

more valuable beyond the observation period. Repeating the<br />

analysis for the six-year customer lifetime value, we show<br />

that referred customers indeed remain more valuable. The<br />

average customer lifetime value of referred customers is<br />

approximately 6 euros higher than that of other customers<br />

(Mann–Whitney test, p < .001). After we control for differences<br />

in customer demographics and time of acquisition,<br />

the value differential is approximately 40 euros (Column 6<br />

of Table 2; p < .001). Because the average customer lifetime<br />

value of a nonreferred customer is 253 euros, a referred customer<br />

is approximately 16% more valuable to the bank than<br />

a comparable nonreferred customer over a horizon of six<br />

years. If we take into account the difference in acquisition<br />

costs of approximately 20 euros, the difference in customer<br />

lifetime value is approximately 25%. This value differential<br />

is quite considerable.<br />

We also assess the extent to which the differences in<br />

customer value are robust across various subsets of customers.<br />

Table 3 reports the regression coefficients for the<br />

referral program in models of customer value, with the<br />

same controls as in the previous analysis in Columns 5 and<br />

6 of Table 2. Row 1 of Table 3 shows that the results for the<br />

retail customer segment are nearly identical to those for the<br />

entire sample. This is not surprising, because retail customers<br />

make up approximately 90% of our overall sample.<br />

More informative is that the difference in customer value<br />

also exists in the nonretail segment (Row 2 of Table 3).<br />

Rows 3 and 4 of Table 3 show that the positive referral<br />

differential exists among high-margin customers, defined as<br />

those in the top decile based on margin, but not low-margin<br />

customers, defined as those in the bottom decile based on<br />

margin. 8 The remaining rows in Table 3 show that sizable<br />

value differentials between referred and nonreferred customers<br />

exist among both men and women and among all<br />

age ranges, except for those over the age of 55. Overall, the<br />

acquisition through a referral program is associated with<br />

higher customer value for the majority of customer types,<br />

but not for all. These results suggest that using referral programs<br />

might not be beneficial in all customer segments, an<br />

idea we develop further in the “Discussion” section.<br />

Robustness Checks<br />

Table 4 shows that the hypothesis tests are robust to including<br />

retail versus nonretail segment membership as an additional<br />

control variable and allowing the effect of the referral<br />

program to vary as a function of age, sex, marital status, and<br />

retail segment membership. Given the results of Table 3, we<br />

8Low-margin customers and high-margin customers are found<br />

in both the retail and the nonretail segments.<br />

TAble 3<br />

Results for Difference in Customer Value Within<br />

Various Segments<br />

Observed Customer<br />

Customer lifetime<br />

Value Value<br />

(Robust (Robust N (N of<br />

Standard Standard Referred<br />

errors) errors) Customers)<br />

Retail customers 48.620*** 39.082*** 8384<br />

(6.574) (6.633) (4473)<br />

Nonretail customers 77.309** 69.803* 1111<br />

(29.855) (30.023) (536)<br />

High margin customers 80.421** 69.669* 950<br />

(27.768) (28.004) (533)<br />

Low margin customers –1.146 –13.212*** 962<br />

(1.581) (2.087) (247)<br />

Male customers 51.679*** 42.305*** 4371<br />

(10.600) (10.669) (2150)<br />

Female customers 47.437*** 38.274*** 5124<br />

(9.604) (9.690) (2859)<br />

65 years of age –1.577 –8.589 1306<br />

*p < .05.<br />

**p < .01.<br />

***p < .001.<br />

(21.421) (21.409) (536)<br />

Notes: Each row displays the coefficient of referral program in models<br />

with the same control variables as in Table 2 but estimated<br />

for specific segments.<br />

also allowed for a nonlinear effect of age. 9 We mean-center<br />

all variables that interact with referral program, so its coefficient<br />

still reflects the main effect. This coefficient keeps<br />

its sign and significance in each model, so the hypotheses<br />

remain supported. The coefficients are larger than in Table<br />

2, in which we did not control for segment membership and<br />

nonlinear age effects, indicating that our main analysis provides<br />

rather conservative estimates of the referral effects.<br />

As a second robustness check, we repeated the analyses<br />

presented in Table 2 for the sample including all outliers.<br />

The direction and significance of the referral effect remained<br />

the same, but the size of several effects increased. The difference<br />

in daily contribution margin increased from 7.6<br />

cents to 16 cents per day, the margin erosion increased from<br />

23.1 cents to 45.4 cents per thousand days, the churn hazard<br />

reduction remained at 20%, and the difference in customer<br />

lifetime value increased from 40 euros to 66 euros. These<br />

9Because some readers may be interested in how the effect of<br />

referral program is moderated by covariates in the time-varying<br />

contribution model, we estimate the latter using a random coefficients<br />

specification rather than the fixed-effects specification used<br />

in Table 2.<br />

Referral Programs and Customer Value / 55


56 / Journal of Marketing, January 2011<br />

TAble 4<br />

Robustness Checks Allowing for Referral effects to be Moderated<br />

H1b H1a Daily<br />

H2b H3 H3 Daily<br />

Contribution<br />

Contribution<br />

Margin (Time<br />

H2a Churn<br />

(Time<br />

Observed<br />

Customer<br />

Customer<br />

lifetime<br />

Margin Varying) Churn Varying) Value Value<br />

Referral program .133*** .228*** –.270** .791 88.413*** 79.695***<br />

(.017) (.062) (.084) (1.479) (10.870) (10.949)<br />

Age (mean centered) a .310*** .529** 1.421*** 1.421*** 199.247*** 164.893***<br />

(.060) (.191) (.361) (.362) (41.286) (41.647)<br />

Age2 (mean centered) a –.004 –.779 –.014 –.014 –2.276 –2.190<br />

(.003) (.733) (.014) (.014) (1.834) (1.848)<br />

Female (mean centered) –.008 .035 .017 .016 –3.822 –2.828<br />

(.013) (.041) (.076) (.076) (9.393) (9.487)<br />

Married (mean centered) .050 .828*** –.445* –.445* 34.372 38.493<br />

(.039) (.126) (.201) (.202) (27.838) (28.187)<br />

Single (mean centered) .080* .933*** –.472* –.473* 52.010 59.890*<br />

(.040) (.128) (.211) (.211) (27.774) (28.119)<br />

Divorced (mean centered) .111* .916*** –.589* –.590* 77.122* 86.686**<br />

(.044) (.139) (.231) (.231) (31.289) (31.696)<br />

Widowed (mean centered) .340*** 1.104*** –1.011*** –1.012*** 236.446*** 254.709***<br />

(.058) (.154) (.272) (.273) (40.821) (41.259)<br />

Nonretail segment (mean centered) .440*** .551*** –.263* –.263* 310.169*** 311.262***<br />

(.031) (.060) (.124) (.124) (21.972) (22.152)<br />

Age ¥ referral programa,b .151 .081 –.531 –.532 99.676 117.447*<br />

(.086) (.255) (.514) (.514) (57.745) (58.183)<br />

Age2 ¥ referral programa, b –.015*** –.311 .022 .022 –10.212*** –10.392***<br />

(.004) (1.018) (.020) (.020) (2.539) (2.556)<br />

Female ¥ referral programb –.005 –.019 –.104 –.104 –2.980 –2.982<br />

(.020) (.056) (.112) (.112) (13.210) (13.320)<br />

Married ¥ referral programb –.162* –.974*** .953** .951** –108.801* –114.852*<br />

(.068) (.174) (.362) (.362) (45.824) (46.049)<br />

Single ¥ referral programb –.097 –.972*** .743* .743* –62.859 –70.679<br />

(.068) (.175) (.366) (.366) (45.988) (46.207)<br />

Divorced ¥ referral programb –.147* –.982*** .916* .917* –103.598* –112.706*<br />

(.074) (.190) (.394) (.394) (50.159) (50.458)<br />

Widowed ¥ referral programb –.270** –1.181*** 1.246** 1.248** –197.134** –211.207**<br />

(.091) (.222) (.456) (.456) (60.695) (61.046)<br />

Nonretail ¥ referral programb –.025 –.084 .014 .013 –49.754 –49.193<br />

(.046) (.084) (.188) (.188) (30.805) (31.001)<br />

Year 2007 (dummy) –1.487***<br />

(.143)<br />

Year 2008 (dummy) –2.562***<br />

(.236)<br />

Cumulative days (in thousands) 4.004***<br />

(.387)<br />

Cumulative days (in thousands) ¥ –.220*<br />

referral program (.101)<br />

Log(duration) ¥ referral program –.167<br />

(.232)<br />

Constant .211*** .244*** 100.219*** 147.265***<br />

(.016) (.056) (9.809) (9.921)<br />

Observations 9495 28,353 9495 9495 9495 9495<br />

R² .107 .099c .123 .122<br />

Log-pseudo-likelihood<br />

*p < .05.<br />

**p < .01.<br />

***p < .001.<br />

aWe divide age by 100 for better readability.<br />

bThe first variable in this interaction is mean centered.<br />

–11,705.8 –11,705.6<br />

cBecause the model is a random coefficients model estimated with residual maximum likelihood, this value is a pseudo-R-square calculated as<br />

the squared correlation between predicted and actual values.<br />

Notes: Robust standard errors are in parentheses. All models include dummies for month of acquisition.


esults suggest that our main analysis is rather conservative<br />

with regard to the size of the referral differentials.<br />

Although hazard analysis properly accounts for right<br />

censoring, managers are also interested in simply knowing<br />

who is likely to have remained with the firm within a certain<br />

time frame. Therefore, we also assessed the relationship<br />

between referral and the probability of still being with<br />

the bank 21 months after acquisition. This time span is the<br />

longest duration observable without right censoring for<br />

every customer, including those who were acquired last, at<br />

the end of December 2006. Using logistic regression and<br />

controlling for demographics and month of acquisition, we<br />

find that referred customers are approximately 22% less<br />

likely to leave the firm within the first 21 months (p < .01).<br />

Computing the customer lifetime value over three,<br />

rather than six, years after acquisition and repeating the<br />

analysis by controlling for demographics and time of acquisition<br />

yields a value differential between referred and nonreferred<br />

customers of 52 euros (p < .001) rather than 40<br />

euros. Both the size and the statistical significance of the<br />

latter value are rather robust to reestimating the model on<br />

smaller random samples of 90% (39 euros, p < .001), 75%<br />

(42 euros, p < .001), 50% (48 euros, p < .001), and 25% (36<br />

euros, p < .01). We also computed the expected value differential<br />

if there were no difference in retention between referred<br />

and nonreferred customers. The differential in six-year customer<br />

lifetime value decreased from 40 euros to 33 euros.<br />

Finally, we extended the model of margin dynamics and<br />

allowed the effect of time and its interaction with referral to<br />

vary as a function of observed customer demographics,<br />

retail versus nonretail status, and normally distributed unobserved<br />

heterogeneity. This extended random coefficients<br />

model did not fit the data better: The squared correlation<br />

between observed and predicted values (pseudo-R 2)<br />

increased by only .1%, and the Bayesian information criterion<br />

even decreased. Most important, the coefficients of<br />

central interest and the statistical inference were not<br />

affected: Customers acquired through referral had a sizable<br />

initial margin advantage that eroded to zero after approximately<br />

1000 days.<br />

Discussion<br />

Key Findings<br />

Evidence of the economic value of stimulated WOM and of<br />

the customers it helps acquire has been sorely lacking. Our<br />

study addresses this gap in the context of referral programs<br />

and documents the attractiveness of customers acquired<br />

through such a program: Contribution margin, retention, and<br />

customer value all were significantly and sizably higher for<br />

referred customers. In short, referred customers are more<br />

valuable in both the short and the long run. Yet we also find<br />

that the effect is not uniform across all types of customers<br />

and that the referral program was less beneficial when used<br />

to acquire older customers or low-margin customers.<br />

In our application, the value of referred customers in the<br />

six years after acquisition was 40 euros (or 16%) higher<br />

than that of nonreferred customers with similar demograph-<br />

ics and time of acquisition. Considering the initial reward of<br />

25 euros given to the referrer as an investment, this implies<br />

a return on investment of approximately 60% over a sixyear<br />

span. This is a conservative estimate because it does not<br />

take into account that the total acquisition costs of referred<br />

customers are approximately 20 euros lower than those of<br />

other customers.<br />

Implications for Practice<br />

Several scholars have expressed cautious skepticism about<br />

the value of “viral-for-hire” and other stimulated WOM<br />

(e.g., Trusov, Bucklin, and Pauwels 2009; Van den Bulte<br />

2010). Doubts about the benefits of stimulated WOM have<br />

long frustrated managers facing demands to increase their<br />

marketing return on investment. Our findings are important<br />

news for practitioners considering deploying customer<br />

referral programs in their own firm. Assuaging prior skepticism,<br />

we document a positive value differential, in both the<br />

short run and the long run, between customers acquired<br />

through a referral program and other customers. Importantly,<br />

this value differential is larger than the referral fee.<br />

Thus, referral programs can indeed pay off.<br />

The positive differential indicates that abuse by opportunistic<br />

customers and other harmful side effects of referral<br />

programs are much less important than their benefits. The<br />

referral program we analyzed was especially prone to<br />

exploitation because no conditions, such as minimum stay<br />

or assets, applied to the newly acquired customer. Finding a<br />

positive value differential of referred customers in this setting<br />

is especially compelling. Moving beyond referral programs<br />

specifically, our study indicates that a stronger focus<br />

on stimulated rather than organic WOM is worth considering<br />

(Godes and Mayzlin 2009).<br />

However, our results also suggest that firms should<br />

think carefully about what prospects to target with referral<br />

programs and how big of a referral fee to provide. For the<br />

program we analyzed, we found that the customer value differential<br />

is much larger in some segments than in others.<br />

People under the age of 55 and high-margin customers are<br />

more attractive to acquire through a referral program. It is<br />

not necessarily a coincidence that these customers also tend<br />

to be more profitable for banks (and many other consumer<br />

marketers). To the extent that the value differential stems<br />

from better matching and social enrichment, as sociological<br />

theory suggests and as employee referral programs have<br />

documented, referral programs do not “create” higher-value<br />

customers by transforming unattractive prospects into<br />

attractive customers. Rather, they help firms selectively<br />

acquire more valuable prospects and retain them longer at<br />

lower cost. Thus, instead of the currently practiced “all-in”<br />

approach, firms should design and target referral programs<br />

such that attractive customers are more likely to be enticed.<br />

Managers must also make their customers aware of their<br />

referral programs. Bank of America, for example, communicates<br />

its referral program on all its automated teller<br />

machines throughout the United States. Connecting referral<br />

programs with online activities might help further increase<br />

their reach beyond existing customers’ networks of strong<br />

Referral Programs and Customer Value / 57


ties and face-to-face interactions. Managers must also make<br />

it convenient for prospects to actually become a customer. A<br />

possible application is to partner with online communities<br />

and make it easy for people to start a relationship with the<br />

firm online, immediately after they receive a referral from<br />

an existing customer in the same community. Our results<br />

suggest that such awareness and facilitation efforts should<br />

be targeted selectively to customers who offer the highest<br />

value differential.<br />

The referral fee is another issue that requires attention<br />

when designing a referral program. Many programs offer<br />

the same reward to each referrer (Kumar, Petersen, and<br />

Leone 2010). Yet, as we show, the value of referred customers<br />

can vary widely even for one company. Thus, firms<br />

may benefit from offering rewards based on the value of the<br />

referred customer. However, the question then becomes<br />

how to do this without adding too much complexity to the<br />

program. There may be a simple answer: A standard<br />

homophily argument suggests that valuable referrers are more<br />

likely to generate valuable referrals. Thus, firms may want to<br />

make the referral fee a function of the value of the referrer.<br />

A different approach to take advantage of the referral<br />

effect would be to try to generate conditions in which nonreferred<br />

customers become subject to the same mechanisms<br />

that make referred customers more valuable. To the extent<br />

that the differences we have documented stem from better<br />

matching, from social enrichment, or from other mechanisms<br />

that firms can actively foster among all customers,<br />

firms may be able to dramatically “scale up” the beneficial<br />

referral effect beyond dyads of referring and referred customers.<br />

For example, pharmaceutical companies increasingly<br />

involve local opinion leaders in their speaker programs<br />

and other medical education efforts. They do so to<br />

capitalize on these physicians’ relevance and credibility<br />

with practicing physicians.<br />

Firms in the same industry often reward referrers with<br />

the same amount. For example, most German banks offer<br />

25 euros for a referral, as does the bank we studied. Our<br />

results indicate that managers set the referral fee rather low,<br />

allowing the firm to reap attractive returns from its program.<br />

Offering higher rewards might lead to even more customer<br />

acquisitions while still providing positive returns on<br />

investment. Firms should calculate the reward considering<br />

their specific program and the customers it attracts instead<br />

of merely following their competitors.<br />

Further Research<br />

Our study focuses on referred and nonreferred customers of<br />

one particular bank. Although its customer base and referral<br />

program have no unusual characteristics, replications would<br />

nonetheless be welcome. Such studies require rich, firmspecific<br />

data on a large set of customers, with individual<br />

profitability observed over a long period. Therefore, we<br />

expect replications and extensions to come from other<br />

single-firm studies such as ours and those of Godes and<br />

Mayzlin (2009), Haenlein (2010), Iyengar, Van den Bulte, and<br />

Valente (2011), and Nitzan and Libai (2010). Because the<br />

mechanisms of better matching and social enrichment are<br />

likely to be more important for complex products with important<br />

experience attributes, rather than simple products with<br />

58 / Journal of Marketing, January 2011<br />

mostly search attributes (e.g., Coverdill 1998; Kornish and<br />

Li 2010; Rees 1966), studies of multiple products with varying<br />

levels of complexity would be especially informative.<br />

It is likely that the quality of the matches with the firm<br />

deteriorates as existing customers refer more new customers.<br />

It would be of practical interest to know at what rate<br />

the quality of referrals decreases and at what point it tends<br />

not to justify the cost of acquisition anymore. It may also be<br />

useful to know if the motivation of the referrer changes<br />

depending on the reward and whether the size of the reward<br />

affects the quality of the referred customer.<br />

Several of the implications for practice point to the<br />

benefits of better understanding the drivers of the value differential<br />

we documented. Although our results are consistent<br />

with the better matching and social enrichment mechanisms<br />

we used to develop our hypotheses, our analysis<br />

focused on the consequences for contribution margin, retention,<br />

and customer value rather than on the intervening<br />

mechanisms. Research aimed at more directly parsing out<br />

the mechanisms is likely to require information about actual<br />

dyads of referring and referred customers. This would<br />

enable researchers to test, for example, the social enrichment<br />

argument by matching the referred customer with the<br />

respective referrer and analyzing the dependence of their<br />

retention. Additional survey data may help document differences<br />

in product knowledge over time and shed light on the<br />

existence of an informational advantage eroding over time.<br />

Having matched dyad-level data on both referring and<br />

referred customers would also make it possible to check<br />

whether referral dyads exhibit homophily and whether the<br />

value of referred customers varies systematically with that<br />

of their referrer (Haenlein 2010; Nitzan and Libai 2010).<br />

This would yield valuable insights for the design of individual<br />

rewards instead of the currently practiced “one-size-fitsall”<br />

approach.<br />

Conclusion<br />

This study provides the first assessment of economically relevant<br />

differences between customers acquired through a referral<br />

program and customers acquired through other methods.<br />

It documents sizable differences in contribution margin,<br />

retention, and customer value; analyzes whether these differences<br />

erode or persist over time; and investigates differences<br />

across customer segments. The finding that, on average,<br />

referred customers are more valuable than other customers<br />

provides the first direct evidence of the financial attractiveness<br />

of referral programs and also offers much-needed evidence<br />

of the financial appeal of stimulated WOM in general.<br />

Improvements in the targeting, design, and implementation<br />

of such programs will require a better understanding<br />

of the drivers of the value differential. The dyadic interdependence<br />

in the behavior of the referrer and the referred<br />

customer deserves special attention in further research<br />

because it is likely to prove highly relevant to both better<br />

theoretical understanding and more effective program<br />

management.


ReFeReNCeS<br />

Belsley, David A., Edwin Kuh, and Roy E. Welsch (1980), Regression<br />

Diagnostics: Identifying Influential Data and Sources of<br />

Collinearity. New York: John Wiley & Sons.<br />

Biggart, Nicole W. (1989), Charismatic Capitalism: Direct Selling<br />

Organizations in America. Chicago: University of Chicago Press.<br />

Biyalogorsky, Eyal, Eitan Gerstner, and Barak Libai (2001), “Customer<br />

Referral Management: Optimal Reward Programs,”<br />

Marketing Science, 20 (1), 82–95.<br />

Blossfeld, Hans-Peter, Alfred Hamerle, and Karl Ulrich Mayer<br />

(1989), Event History Analysis: Statistical Theory and Application<br />

in the Social Sciences. Hillsdale, NJ: Lawrence Erlbaum<br />

Associates.<br />

Bolton, Patrick, Xavier Freixas, and Joel Shapiro (2007), “Conflicts<br />

of Interest, Information Provision, and Competition in the<br />

Financial Services Industry,” Journal of Financial Economics,<br />

85 (2), 297–330.<br />

Castilla, Emilio J. (2005), “Social Networks and Employee Performance<br />

in a Call Center,” American Journal of Sociology, 110<br />

(5), 1243–83.<br />

Choi, Jeonghye (2009), “Social Influence from Existing to New<br />

Customers,” working paper, the Wharton School, University of<br />

Pennsylvania.<br />

Coverdill, James E. (1998), “Personal Contacts and Post-Hire Outcomes:<br />

Theoretical and Empirical Notes on the Significance of<br />

Matching Methods,” Research in Social Stratification and<br />

Mobility, 16, 247–69.<br />

Edwards, Helen and Derek Day (2005), Creating Passion Brands:<br />

Why It’s Time to Lead from the Heart. London: Kogan Page.<br />

Fernandez, Roberto M., Emilio J. Castilla, and Paul Moore (2000),<br />

“Social Capital at Work: Networks and Employment at a Phone<br />

Center,” American Journal of Sociology, 105 (5), 1288–1356.<br />

Fleming, John H., Curt Coffman, and James K. Harter (2005),<br />

“Manage Your Human Sigma,” Harvard Business Review, 83<br />

(7–8), 106–114.<br />

Godes, David and Dina Mayzlin (2009), “Firm-Created Word-of-<br />

Mouth Communication: Evidence from a Field Test,” Marketing<br />

Science, 28 (4), 721–39.<br />

Gouldner, Alvin W. (1960), “The Norm of Reciprocity: A Preliminary<br />

Statement,” American Sociological Review, 25 (April), 161–78.<br />

Gupta, Sunil and Valarie Zeithaml (2006), “Customer Metrics and<br />

Their Impact on Financial Performance,” Marketing Science,<br />

25 (6), 718–39.<br />

Guseva, Alya (2008), Into the Red: The Birth of the Credit Card<br />

Market in Postcommunist Russia. Palo Alto, CA: Stanford University<br />

Press.<br />

Haenlein, Michael (2010), “A Social Network Analysis of Customer-<br />

Level Revenue Distribution,” Marketing Letters, forthcoming.<br />

Helm, Sabrina (2003), “Calculating the Value of Customers’<br />

Referrals,” Managing Service Quality, 13 (2), 124–33.<br />

Irwin, Julie R. and Gary H. McClelland (2001), “Misleading<br />

Heuristics and Moderated Multiple Regression Models,” Journal<br />

of Marketing Research, 38 (February), 100–109.<br />

Iyengar, Raghuram, Christophe Van den Bulte, and Thomas W.<br />

Valente (2011), “Opinion Leadership and Social Contagion in<br />

New Product Diffusion,” Marketing Science, forthcoming.<br />

Kornish, Laura J. and Quiping Li (2010), “Optimal Referral<br />

Bonuses with Asymmetric Information: Firm-Offered and<br />

Interpersonal Incentives,” Marketing Science, 29 (1), 108–121.<br />

Kumar, V., J. Andrew Petersen, and Robert P. Leone (2007), “How<br />

Valuable Is Word of Mouth?” Harvard Business Review, 85<br />

(10), 139–46.<br />

———, ———, and ——— (2010), “<strong>Dr</strong>iving <strong>Prof</strong>itability by<br />

Encouraging Customer Referrals: Who, When, and How,”<br />

Journal of Marketing, 74 (September), 1–17.<br />

——— and Denish Shah (2009), “Expanding the Role of Marketing:<br />

From Customer Equity to Market Capitalization,” Journal<br />

of Marketing, 73 (November), 119–36.<br />

Lewis, Michael (2006), “Customer Acquisition Promotions and<br />

Customer Asset Value,” Journal of Marketing Research, 43<br />

(May), 195–203.<br />

Meyer, Bruce D. (1990), “Unemployment Insurance and Unemployment<br />

Spells,” Econometrica, 58 (July), 757–82.<br />

Montgomery, James D. (1991). “Social Networks and <strong>Lab</strong>or-Market<br />

Outcomes: Toward an Economic Analysis,” American Economic<br />

Review, 81 (December), 1408–1418.<br />

Mummert, Hallie (2000), “The Year’s Best Bells & Whistles,” Target<br />

Marketing, 23 (11), 3–5.<br />

Neckerman, Kathryn M. and Roberto M. Fernandez (2003),<br />

“Keeping a Job: Network Hiring and Turnover in a Retail<br />

Bank,” Research in the Sociology of Organizations, 20,<br />

299–318.<br />

Nitzan, Irit and Barak Libai (2010), “Social Effects on Customer<br />

Retention,” Report No. 10-107. Cambridge, MA: Marketing<br />

Science Institute.<br />

Parasuraman, A., Valarie A. Zeithaml, and Leonard L. Berry<br />

(1985), “A Conceptual Model of Service Quality and Its Implications<br />

for Future Research,” Journal of Marketing, 49 (Fall),<br />

41–50.<br />

Rees, Albert (1966), “Information Networks in <strong>Lab</strong>or Markets,”<br />

American Economic Review, 56 (March), 559–66.<br />

Ryu, Gangseog and Lawrence Feick (2007), “A Penny for Your<br />

Thoughts: Referral Reward Programs and Referral Likelihood,”<br />

Journal of Marketing, 71 (January), 84–94.<br />

Svensson, Lars E. (1994), “Estimating and Interpreting Forward<br />

Interest Rates: Sweden 1992-1994,” NBER Working Paper<br />

No. 4871. Cambridge, MA: National Bureau of Economic<br />

Research.<br />

Thaler, Richard H. and Cass R. Sunstein (2008), Nudge: Improving<br />

Decisions About Health, Wealth, and Happiness. New<br />

Haven, CT: Yale University Press.<br />

Trusov, Michael, Randolph E. Bucklin, and Koen Pauwels (2009),<br />

“Effects of Word-of-Mouth Versus Traditional Marketing:<br />

Findings from an Internet Social Networking Site,” Journal of<br />

Marketing, 73 (September), 90–102.<br />

Van den Bulte, Christophe (2010), “Opportunities and Challenges<br />

in Studying Customer Networks,” in The Connected Customer:<br />

The Changing Nature of Consumer and Business Markets, S.<br />

Wuyts, M.G. Dekimpe, E. Gijsbrechts, and R. Pieters, eds.<br />

London: Routledge, 7–35.<br />

——— and Stefan Wuyts (2007), Social Networks and Marketing.<br />

Cambridge, MA: Marketing Science Institute.<br />

Villanueva, Julian, Shijin Yoo, and Dominique M. Hanssens<br />

(2008), “The Impact of Marketing-Induced Versus Word-of-<br />

Mouth Customer Acquisition on Customer Equity Growth,”<br />

Journal of Marketing Research, 45 (February), 48–59.<br />

Wiesel, Thorsten, Bernd Skiera, and Julian Villanueva (2008),<br />

“Customer Equity: An Integral Part of Financial Reporting,”<br />

Journal of Marketing, 72 (March), 1–14.<br />

Wirtz, Jochen and Patricia Chew (2002), “The Effects of Incentives,<br />

Deal Proneness, Satisfaction and Tie Strength on Wordof-Mouth<br />

Behaviour,” International Journal of Service Industry<br />

Management, 13 (2), 141–62.<br />

Wooldridge, Jeffrey M. (2002), Introductory Econometrics: A<br />

Modern Approach, 2d ed. Cincinnati: South-Western.<br />

Yim, Chi Kin, David K. Tse, and Kimmy Wa Chan (2008),<br />

“Strengthening Customer Loyalty Through Intimacy and Passion:<br />

Roles of Customer–Firm Affection and Customer–Staff<br />

Relationships in Services,” Journal of Marketing Research, 45<br />

(December), 741–56.<br />

ZatzNotFunny.com (2008), “Confirmed: TiVo Rewards Killed,”<br />

(accessed June 28, 2010), [available at http://www.zatznotfunny.<br />

com/2008-02/confirmed-tivo-rewards-killed/].<br />

Referral Programs and Customer Value / 59


Oliver Hinz, Bernd Skiera, Christian Barrot, & Jan U. Becker<br />

Seeding Strategies for Viral<br />

Marketing: An Empirical<br />

Comparison<br />

Seeding strategies have strong influences on the success of viral marketing campaigns, but previous studies<br />

using computer simulations and analytical models have produced conflicting recommendations about the optimal<br />

seeding strategy. This study compares four seeding strategies in two complementary small-scale field experiments,<br />

as well as in one real-life viral marketing campaign involving more than 200,000 customers of a mobile phone<br />

service provider. The empirical results show that the best seeding strategies can be up to eight times more<br />

successful than other seeding strategies. Seeding to well-connected people is the most successful approach<br />

because these attractive seeding points are more likely to participate in viral marketing campaigns. This finding<br />

contradicts a common assumption in other studies. Well-connected people also actively use their greater reach<br />

but do not have more influence on their peers than do less well-connected people.<br />

Keywords: viral marketing, seeding strategy, word of mouth, social contagion, targeting<br />

The future of traditional massmedia advertising is<br />

uncertain in the modern environment of increasingly<br />

prevalent digital video recorders and spam filters.<br />

Marketers must realize that 65% of consumers consider<br />

themselves overwhelmed by too many advertising messages,<br />

and nearly 60% believe advertising is not relevant<br />

to them (Porter and Golan 2006). Such information overload<br />

can cause consumers to defer their purchase altogether<br />

(Iyengar and Lepper 2000), and strong evidence indicates<br />

consumers actively avoid traditional marketing instruments<br />

(Hann et al. 2008).<br />

Other empirical evidence also reveals that consumers<br />

increasingly rely on advice from others in personal or professional<br />

networks when making purchase decisions (Hill,<br />

Provost, and Volinsky 2006; Iyengar, Van den Bulte, and<br />

Valente 2011; Schmitt, Skiera, and Van den Bulte 2011).<br />

In particular, online communication appears increasingly<br />

important as more websites offer user-generated content,<br />

Oliver Hinz is Chaired <strong>Prof</strong>essor of Information Systems esp.<br />

Electronic Markets, Technische Universität Darmstadt (e-mail: hinz<br />

@wi.tu-darmstadt.de). Bernd Skiera is Chaired <strong>Prof</strong>essor of Electronic<br />

Commerce, Department of Marketing, Goethe-University of<br />

Frankfurt (e-mail: skiera@wiwi.uni-frankfurt.de). Christian Barrot is<br />

Assistant <strong>Prof</strong>essor of Marketing and Innovation (e-mail: christian<br />

.barrot@the-klu.org), and Jan U. Becker is Assistant <strong>Prof</strong>essor<br />

of Marketing and Service Management (e-mail: jan.becker@theklu.org),<br />

Kühne Logistics University. The authors thank Carsten<br />

Takac, Philipp Schmitt, Martin Spann, Lucas Bremer, Christian<br />

Messerschmidt, Daniele Mahoutchian, Nadine Schmidt, and Katharina<br />

Schnell for helpful input on previous versions of this article. In<br />

addition, three anonymous JM referees provided many helpful suggestions.<br />

This research is supported by the E-<strong>Finance</strong> <strong>Lab</strong> Frankfurt.<br />

© 2011, American Marketing Association<br />

ISSN: 0022-2429 (print), 1547-7185 (electronic) 55<br />

such as blogs, video and photo sharing opportunities,<br />

and online social networking platforms (e.g., Facebook,<br />

LinkedIn). Companies have adapted to these trends by<br />

shifting their budgets from above-the-line (mass media) to<br />

below-the-line (e.g., promotions, direct mail, viral) marketing<br />

activities.<br />

Not surprisingly then, viral marketing has become a<br />

hot topic. The term “viral marketing” describes the phenomenon<br />

by which consumers mutually share and spread<br />

marketing-relevant information, initially sent out deliberately<br />

by marketers to stimulate and capitalize on word-ofmouth<br />

(WOM) behaviors (Van der Lans et al. 2010). Such<br />

stimuli, often in the form of e-mails, are usually unsolicited<br />

(De Bruyn and Lilien 2008) but easily forwarded to multiple<br />

recipients. These characteristics parallel the traits of<br />

infectious diseases, such that the name and many conceptual<br />

ideas underlying viral marketing build on findings from<br />

epidemiology (Watts and Peretti 2007).<br />

Because viral marketing campaigns leave the dispersion<br />

of marketing messages up to consumers, they tend to be<br />

more cost efficient than traditional massmedia advertising.<br />

For example, one of the first successful viral campaigns,<br />

conducted by Hotmail, generated 12 million subscribers in<br />

just 18 months with a marketing budget of only $50,000.<br />

Google’s Gmail captured a significant share of the e-mail<br />

provider market, even though the only way to sign up for<br />

the service was through a referral. A recent viral advertisement<br />

by Tipp-Ex (“A hunter shoots a bear!”) triggered<br />

nearly 10 million clicks in just four weeks.<br />

However, to enjoy such results, firms must consider four<br />

critical viral marketing success factors: (1) content, in that<br />

the attractiveness of a message makes it memorable (Berger<br />

and Milkman 2011; Berger and Schwartz 2011; Gladwell<br />

2002; Porter and Golan 2006); (2) the structure of the social<br />

Journal of Marketing<br />

Vol. 75 (November 2011), 55–71


network (Bampo et al. 2008); (3) the behavioral characteristics<br />

of the recipients and their incentives for sharing the<br />

message (Arndt 1967); and (4) the seeding strategy, which<br />

determines the initial set of targeted consumers chosen by<br />

the initiator of the viral marketing campaign (Bampo et al.<br />

2008; Kalish, Mahajan, and Muller 1995; Libai, Muller,<br />

and Peres 2005). This last factor is of particular importance<br />

because it falls entirely under the control of the initiator<br />

and can exploit social characteristics (Toubia, Stephen, and<br />

Freud 2010) or observable network metrics. Unfortunately,<br />

there is a “need for more sophisticated and targeted seeding<br />

experimentation” to gain “a better understanding of the role<br />

of hubs in seeding strategies” (Bampo et al. 2008, p. 289).<br />

The conventional wisdom adopts the influentials hypothesis,<br />

which states that targeting opinion leaders and<br />

strongly connected members of social networks (i.e., hubs)<br />

ensures rapid diffusion (for a summary of arguments, see<br />

Iyengar, Van den Bulte, and Valente 2011). However, recent<br />

findings raise doubts. Van den Bulte and Lilien (2001)<br />

show that social contagion, which occurs when adoption is<br />

a function of exposure to other people’s knowledge, attitudes,<br />

or behaviors (Van den Bulte and Wuyts 2007), does<br />

not necessarily influence diffusion, and yet it remains a<br />

basic premise of viral marketing. Such contagion frequently<br />

arises when people who are close in the social structure use<br />

one another to manage uncertainty in prospective decisions<br />

(Granovetter 1985). However, in a computer simulation,<br />

Watts and Dodds (2007) show that well-connected people<br />

are less important as initiators of large cascades of referrals<br />

or early adopters. Their finding, which Thompson (2008)<br />

provocatively summarizes by implying “the tipping point is<br />

toast,” has stimulated a heated debate about optimal seeding<br />

strategies, though no research offers an extensive empirical<br />

comparison of seeding strategies. Van den Bulte (2010)<br />

thus calls for empirical comparisons of seeding strategies<br />

that use sociometric measures (i.e., metrics that capture the<br />

social position of people).<br />

In response, we undertake an empirical comparison of<br />

the success of different seeding strategies for viral marketing<br />

campaigns and identify reasons for variations in these<br />

levels of success. In doing so, we determine whether companies<br />

should care about the seeding of their viral marketing<br />

campaigns and why. In particular, we study whether<br />

well-connected people really are more difficult to activate,<br />

participate more actively in viral campaigns, and have more<br />

influence on their peers than less well-connected people. In<br />

contrast with previous studies that rely on analytical models<br />

or computer simulations, we derive our results from field<br />

experiments and from a real-life viral marketing campaign.<br />

We begin this article by presenting literature relevant to<br />

viral marketing and social contagion theory. We introduce<br />

our theoretical framework, which disentangles the determinants<br />

of social contagion, and present four seeding strategies.<br />

Next, we empirically compare the success of these<br />

seeding strategies in two complementary field experiments<br />

(Studies 1 and 2) that aim at spreading information and<br />

inducing attitudinal changes. Then, we analyze a real-life<br />

viral marketing campaign designed to increase sales, which<br />

provides an economic measure of success. After we identify<br />

the determinants of success, we conclude with a discussion<br />

of our research contributions, managerial implications,<br />

and limitations.<br />

56 / Journal of Marketing, November 2011<br />

Theoretical Framework<br />

When information about an underlying social network is<br />

available, seeding on the basis of this information, as<br />

typically captured by sociometric data, seems promising<br />

(Van den Bulte 2010). Such a strategy can distinguish three<br />

types of people: “hubs,” who are well-connected people<br />

with a high number of connections to others; “fringes,” who<br />

are poorly connected; and “bridges,” who connect two otherwise<br />

unconnected parts of the network. The sociometric<br />

measure of degree centrality captures connectedness within<br />

the local environment (for details, see the Web Appendix at<br />

http://www.marketingpower.com/jmnov11), such that highdegree-centrality<br />

values characterize hubs, whereas lowdegree-centrality<br />

values mark fringes. In contrast, the<br />

sociometric betweenness centrality measure describes the<br />

extent to which a person acts as a network intermediary,<br />

according to the share of shortest communication paths<br />

that pass through that person (see the Web Appendix).<br />

Thus, bridges earn high values on betweenness centrality<br />

measures.<br />

Determinants of Social Contagion<br />

Following Van der Lans et al. (2010), we propose a fourdeterminant<br />

model of social contagion to determine the<br />

success of viral marketing campaigns. First, individual i<br />

receives a viral message from sender s, who can be either a<br />

friend or the campaign initiator that makes i aware of and<br />

informed by the message with information probability I i.<br />

Individual i then may become active and participate in the<br />

campaign with participation probability P i. Given participation,<br />

individual i passes the message to a set of recipients J i,<br />

where n i is the number of recipients (�J i� = n i), such that it<br />

provides a measure of used reach. The number of expected<br />

referrals R i by individual i, then, is the product of the information<br />

probability (I i), the probability of participating (P i),<br />

and the used reach (n i): R i = I i × P i × n i.<br />

The conversion rate w i�j linearly influences the number of<br />

expected successful referrals SR i of individual i on recipients<br />

j (j ∈ J i), given by<br />

(1)<br />

n �i<br />

SRi = Ii × Pi × ni ×<br />

j = 1<br />

w i� j<br />

If a sender i has the same conversion rate for all recipients,<br />

such that w i�j = w i ∀ j ∈ J i, the number of expected<br />

successful referrals can be rewritten as<br />

(2)<br />

n i<br />

SR i = I i × P i × n i × w i�<br />

All these determinants are a function of i’s social position,<br />

though they also may be influenced by the characteristics<br />

of the sender s and the conversion rate w s. Despite<br />

the lack of empirical comparisons of seeding strategies for<br />

viral marketing campaigns, various studies in marketing,<br />

sociology, and epidemiology have analyzed the influence<br />

of the social position (captured by sociometric measures)<br />

on different determinants, such as whether hubs are more<br />

likely to persuade their peers. In Table 1, we summarize<br />

these findings according to the determinants of information<br />

probability I i, participation probability P i, used reach n i,<br />

and conversion rate w i.<br />


TABLE 1<br />

Previous Research<br />

Social Position Has Positive Influence on � � �<br />

Expected Empirically<br />

Expected number of Recommendation Tested<br />

Reason for Participation Used Number of Conversion Successfull for Optimal Seeding<br />

Studies Context Contagion Probability P i Reach n i Referrals R i Rate w i SR i Seeding Strategy Strategy<br />

Coleman, Katz, and Product (low risk) A, BU, NP Hub Hub Hub<br />

Menzel (1966)<br />

Becker (1970) Product (low risk) A, BU, NP Hub Hub Hub<br />

Product (high risk) Fringe Fringe Fringe<br />

Simmel (1950); Porter Messages A Fringe Fringe<br />

and Donthu (2008) Messages A<br />

Watts and Dodds (2007) — — Fringe Hub Fringe Fringe Fringe<br />

Leskovec, Adamic, Product (low risk) A, BU Hub Hub Hub Fringe<br />

and Huberman (2007)<br />

Anderson and May Epidemiology A Hub Hub Hub Hub<br />

(1991); Kemper (1980) Epidemiology A<br />

Granovetter (1973); Messages A Bridge Bridge Bridge<br />

Rayport (1996) Messages A<br />

Iyengar, Van den Bulte, Product (high risk) A, BU Hub Hub Hub Hub<br />

and Valente (2011)<br />

Study 1 Messages A Controlled � Hub, fringe, bridge,<br />

random<br />

Study 2 Messages A � Hub, fringe, bridge,<br />

random<br />

Study 3 Product (low risk) A, BU � � � � � Hub, fringe, random<br />

Notes: A = awareness, BU = belief updating, NP = normative pressure, and i = focal individual. Expected number of referrals: R i = P i × n i; Successful number of referrals: SR i = w i × R i.<br />

Seeding Strategies for Viral Marketing / 57


Effect of social position on information and participation<br />

probability. A viral marketing campaign aims to<br />

inform consumers about the viral marketing message and<br />

to encourage them to participate in the campaign by sending<br />

the message to others. In investigating the impact<br />

of social position on information probability, Goldenberg<br />

et al. (2009) indicate that hubs tend to be better informed<br />

than others because they are exposed to innovations earlier<br />

through their multiple social links. In his reanalysis of<br />

Coleman, Katz, and Menzel’s (1966) medical innovation<br />

study, Burt (1987) also recognizes that some people experience<br />

discomfort when peers whose approval they value<br />

adopt an innovation they have not yet adopted; in this case,<br />

social contagion (reflected in a greater probability to participate)<br />

results from normative pressure and status considerations.<br />

This mechanism could explain why Coleman, Katz,<br />

and Menzel (1966) find that highly integrated people (e.g.,<br />

hubs) are more likely to adopt an innovation early than are<br />

more isolated people.<br />

However, in some cases, hubs do not adopt innovations<br />

first (Becker 1970), such as when the innovation does not<br />

suit the hub’s opinion, which may mean that adoption<br />

occurs first at the fringes of the network (Iyengar, Van den<br />

Bulte, and Valente 2011). Another potential explanation of<br />

hubs’ lower participation probability stems from information<br />

overload effects. Because hubs are exposed to so many<br />

contacts, they possess a wealth of information and thus<br />

might be more difficult to activate (Porter and Donthu 2008;<br />

Simmel 1950) or less likely to participate in viral marketing<br />

campaigns. Overall, information and participation probabilities<br />

remain difficult to disentangle; for the purposes of<br />

this study, we assume that all receivers of viral marketing<br />

messages are aware of them. This assumption is likely to<br />

hold for our three empirical studies, and our main findings<br />

remain unchanged even when it does not. The only difference<br />

is that the participation probability would also capture<br />

the probability that a person is informed (and thus aware)<br />

of the viral marketing campaign.<br />

Effect of social position on used reach. Epidemiology<br />

studies indicate that hub constellations foster the spread of<br />

diseases (Anderson and May 1991; Kemper 1980), which<br />

suggests in parallel that hubs should be more attractive for<br />

seeding viral marketing campaigns. However, it is unclear<br />

whether hubs actively and purposefully make use of their<br />

potential reach. The deliberate use of reach is a common<br />

assumption, and yet only Leskovec, Adamic, and<br />

Huberman (2007) actually confirm that hubs send more<br />

messages. Furthermore, their definition of hubs relies on<br />

messaging behavior, such that it cannot offer generalizable<br />

evidence of the assumption that hubs actively use their<br />

greater reach potential.<br />

In contrast, we anticipate that individual i’s used reach<br />

(first generation), added to the used reach of successive<br />

generations that originate from i’s initial direct reach (second<br />

and further generations), which we call i’s “influence<br />

domain” (Lin 1976), depends on the number of others who<br />

already have received the message. In this setting, bridges<br />

are advantageous because they can forward the message to<br />

different parts of the network (Granovetter 1973) that have<br />

not yet been infected with the viral campaign.<br />

58 / Journal of Marketing, November 2011<br />

Effect of social position on conversion rate. A person’s<br />

social position might also indicate the degree of persuasiveness,<br />

as measured by the conversion rate—namely, the<br />

share of referrals that lead to successful referrals. Iyengar,<br />

Van den Bulte, and Valente (2011) find that hubs are more<br />

likely to be heavy users and that, therefore, their influence<br />

is more effective, because they act in accordance with<br />

their own recommendations (e.g., by making heavy use of<br />

the innovation). Leskovec, Adamic, and Huberman (2007)<br />

find that the success rate per recommendation decreases<br />

with the number of recommendations made, which implies<br />

that people have influence over a limited number of friends<br />

but not over everybody they know. This result indicates<br />

that the conversion rate decreases when hubs use their full<br />

reach potential, though it does not preclude the notion that<br />

hubs instead might select relevant subsets of recipients from<br />

among their peers and thus achieve high conversion rates.<br />

Thus, the effect of social position on the conversion rate<br />

remains unclear. Goldenberg et al. (2009) still make what<br />

they call the conservative assumption that hubs are not more<br />

persuasive than others, though without empirical support.<br />

Seeding Strategies<br />

As our review in Table 1 shows, little consensus exists<br />

regarding recommendations for optimal seeding strategies.<br />

Four studies recommend seeding hubs, three recommend<br />

fringes, and one recommends bridges. We analyze these<br />

discrepant recommendations in turn.<br />

If at least one of the determinants I i, P i, n i, or w i increases<br />

with the connectivity of the sender i, and the<br />

remaining determinants are not correlated with higher connectivity,<br />

hubs should be the targets of initial seeding<br />

efforts, because they spread viral information best—as indicated<br />

in Hanaki et al. (2007), Van den Bulte and Joshi<br />

(2007), and Kiss and Bichler (2008). Using hubs as initial<br />

seeding points implies a “high-degree seeding” strategy.<br />

In contrast, Watts and Dodds’s (2007) computer simulations<br />

of interpersonal influence indicate that targeting<br />

well-connected hubs to maximize the spread of information<br />

works only under certain conditions and may be the exception<br />

rather than the rule. They propose instead that a critical<br />

mass of influenceable people, rather than particularly influential<br />

people, drives cascades of influence. The impact on<br />

triggering critical mass is not even proportional to the number<br />

of people who hubs directly influence; rather, according<br />

to Dodds and Watts (2004), the people most easily influenced<br />

have the highest impact on the diffusion. Moreover, if<br />

hubs suffer from information overload because of their central<br />

position in the social network (Porter and Donthu 2008;<br />

Simmel 1950), they must filter or validate the vast amount<br />

of information they receive, such that they may be less<br />

susceptible to information received from anyone outside<br />

their trusted network. In their analytical model, Galeotti<br />

and Goyal (2009) propose targeting low-degree members<br />

instead, on the fringe of the network, if the probability of<br />

adopting a product increases with the absolute number of<br />

adopting neighbors. Sundararajan (2006) similarly suggests<br />

seeding the fringes rather than the hubs, which we refer to<br />

as a “low-degree seeding” strategy.<br />

When analyses focus on the influence domain, encompassing<br />

referrals beyond the first generation, it also<br />

becomes necessary to consider centrality beyond the local


environment. Bridges who connect otherwise separated<br />

subnetworks have vast influence domains, such that seeding<br />

them might enable information to diffuse throughout parts<br />

of the network and prevent a viral message from simply<br />

circulating in an already infected, highly clustered subnetwork.<br />

Accordingly, Rayport (1996) recommends exploiting<br />

the strength of weak ties (i.e., bridges; Granovetter 1973)<br />

to spread a marketing virus. From an opposite perspective,<br />

Watts (2004) similarly recommends eliminating bridges to<br />

prevent epidemics. We thus refer to the idea of seeding<br />

bridges as a “high-betweenness seeding” strategy.<br />

Finally, if there is no correlation between social position<br />

and the determinants I i, P i, n i, and w i, or if the opposing<br />

influences of the determinants nullify one another, there<br />

should be no differences across the proposed strategies or<br />

a random targeting. We also test this “random seeding”<br />

strategy, which further serves as a benchmark situation in<br />

which no information about the social network is available.<br />

Method<br />

We use three studies to empirically compare the success of<br />

seeding strategies and identify which of the determinants are<br />

influenced by the people’s social position. Our three studies<br />

encompass two types of settings that are particularly relevant<br />

for viral marketing. First, viral marketing campaigns<br />

primarily aim to spread information, create awareness,<br />

and improve brand perceptions, which are noneconomic<br />

goals. Second, other campaigns attempt to increase sales<br />

through mutual information exchanges between adopters<br />

and prospective adopters, to trigger belief updating, such<br />

that we can use an economic measure of success.<br />

These goals map well onto the classification that Van den<br />

Bulte and Wuyts (2007) provide to describe five reasons<br />

for social contagion, the first two of which are especially<br />

relevant for viral marketing campaigns. First, people may<br />

become aware of the existence of an innovation through<br />

WOM provided by previous adopters in a simple information<br />

transfer. Second, people may update their beliefs about<br />

the benefits and costs of a product or service. Third, social<br />

contagion may occur through normative pressures, such<br />

that people experience discomfort when they do not comply<br />

with the expectations of their peer group. Fourth, social<br />

contagion can be based on status considerations and competitive<br />

concerns (i.e., the level of competitiveness between<br />

two people). Fifth, complementary network effects might<br />

cause social contagion, in which the benefit of using a product<br />

or service increases with the number of users.<br />

To examine both types of viral marketing campaigns, we<br />

conduct two experimental studies (Studies 1 and 2) that<br />

simulate viral marketing campaigns in which social contagion<br />

mainly involves simple information transfers and<br />

results in greater awareness as a noneconomic measure of<br />

success. The aim is to compare the success of different<br />

seeding strategies. In Study 3, we examine a viral marketing<br />

campaign in which social contagion relies on belief<br />

updating and results in sales (i.e., economic measure). In<br />

Table 2, we summarize the complementary setup of the<br />

three studies, which helps us overcome some individual<br />

limitations of each study.<br />

Experimental Comparison of Seeding Strategies<br />

In Studies 1 and 2, we compare the success of our four<br />

seeding strategies in different conditions and confirm the<br />

robustness of the results across different settings. Trusov,<br />

Bodapati, and Bucklin (2010) point out the neccessity to<br />

conduct such experiments: In analyzing data from a major<br />

social networking site, they find that only approximately<br />

one-fifth of a user’s friends actually influence that user’s<br />

activity on the site. However, they cannot discern how<br />

responsive the “top influencers” are or whether marketers<br />

should use information about underlying social networks to<br />

seed their viral marketing campaigns. Therefore, they call<br />

for further research that uses straightforward field experiments.<br />

Because such experiments can help identify bestpractice<br />

strategies, we compare the four seeding strategies<br />

in two small-scale field experiments.<br />

Study 1: Comparison of seeding strategies in a controlled<br />

setting. We begin with a controlled setup to ensure internal<br />

validity and control for willingness to actively participate P i<br />

(see Table 1). We recruited 120 students from a German<br />

university. The recruitment and commitment processes<br />

ensured relatively similar participants in terms of communication<br />

activity across treatments because all of them<br />

expressed a willingness to contribute actively. Therefore,<br />

we expect minimal variation in activity levels, compared<br />

with a study in which respondents are unaware of their participation<br />

or do not come into direct contact with the experimenter.<br />

A prerequisite for participation was maintaining<br />

an account on a specified online social networking platform<br />

(similar to Facebook). Using proprietary software, we<br />

automatically gathered each participant’s friends list from<br />

the platform and then applied an event-based approach<br />

to specify boundaries, such that we discarded all links<br />

to friends who did not participate in the experiment.<br />

The software Pajek calculated the sociometric measures<br />

(degree centrality and betweenness centrality; see the Web<br />

Appendix at http://www.marketingpower.com/jmnov11) for<br />

each participant.<br />

The social network thus generated consisted of 120<br />

nodes (i.e., participants) with 270 edges (i.e., friendship<br />

relations). Degree centrality ranged from 1 to 17, with a<br />

mean of 4.463 and a standard deviation of 3.362. In other<br />

words, the participants had slightly more than four friends<br />

each, on average, in the respective, bounded social network.<br />

The correlation (.592, p < �01) between the degree centrality<br />

in this small, bounded network created by the artificial<br />

boundary specification strategy (using the criteria “participation<br />

in experiment”) and the degree centrality of the<br />

entire network hosted by this social networking platform<br />

(6.2 million unique users, November 2009) is striking. It<br />

also supports Costenbader and Valente’s (2003) claim that<br />

some centrality metrics are relatively robust across different<br />

network boundaries. That is, the boundary we applied<br />

does not appear to bias degree centrality, even for a sub<br />

sample that comprises as little as .002% of the entire social<br />

network. The betweenness centrality ranged from 0 to .320,<br />

with a standard deviation of .053.<br />

We used these sociometric measures to implement our<br />

four seeding strategies. The seeding relied on the message<br />

function of the social networking platform, such that<br />

we sent unique tokens of information to a varying subset<br />

of participants (the total population remained unchanged<br />

throughout this experiment) and traced the contagion pro-<br />

Seeding Strategies for Viral Marketing / 59


TABLE 2<br />

Summary of Studies<br />

Study 1 Study 2 Study 3<br />

Seeding strategies Four seeding strategies: Four seeding strategies: Three seeding strategies:<br />

•High degree (HD) •High degree (HD) •High degree (HD)<br />

•Low degree (LD) •Low degree (LD) •Low degree (LD)<br />

•High betweenness (HB) •High betweenness (HB) •Random<br />

•Random (control) •Random (control)<br />

Social contagion through Awareness (advertisement) Awareness (advertisement) Belief updating (service<br />

referral)<br />

Motivation Extrinsic motivation for<br />

sharing (experimental<br />

remuneration)<br />

Intrinsic motivation (funny<br />

video about university)<br />

Extrinsic motivation for<br />

sharing (additional airtime for<br />

referral)<br />

Seeding size 10% of network size 7% of network size Entire network<br />

20% of network size<br />

Seeding timing Sequential Parallel —<br />

Social network 120 nodes (small network),<br />

270 edges<br />

1380 nodes (medium-sized<br />

network), 4052 edges<br />

208,829 nodes (very large<br />

network), 7,786,019 edges<br />

Number of treatments 16 = 4 × 2 × 2 4 —<br />

Number of replications 2 (4 treatments missing) 1 —<br />

Number of experimental<br />

settings 28 4 —<br />

Boundary of network Artificial Natural Natural<br />

Design strengths Test of causality, strong<br />

control due to experimental<br />

setup, identification of<br />

individual behavior due to<br />

specific IDs<br />

Design weaknesses Repeated measures due to<br />

sequential timing, artificial<br />

scenario<br />

Specific finding HD and HB are comparable<br />

and outperform random by<br />

+39%–52% and LD by<br />

factor 7–8<br />

cess. These tokens were to be shared by initial recipients<br />

with friends, who in turn were to spread them further. All<br />

receivers were asked to enter the tokens on a website that<br />

we created for this purpose, along with details about from<br />

whom they received these tokens (called the “referrer”).<br />

Because each participant was provided with unique login<br />

information for this website, we could observe the number<br />

of tokens entered on the website (and thus the number of<br />

successful referrals SR i) by each individual i for each of<br />

the seeding strategies. Furthermore, we could distinguish<br />

whether the recipient received the tokens directly from the<br />

experimenter (“Seeded by Experimenter”) or through viral<br />

spreading from friends. We prohibited and did not observe<br />

the use of forums or mailing lists to spread the tokens.<br />

The experiment used a 4 × 2 × 2 full-factorial design.<br />

Following the strategies we defined previously, we seeded<br />

the tokens every few days to hubs (high-degree seeding),<br />

fringes (low-degree seeding), bridges (high-betweenness<br />

seeding), or a random set of participants. We varied the<br />

number of initial seeds, such that the tokens were sent<br />

to either 12 (10%) or 24 (20%) of the 120 participants.<br />

60 / Journal of Marketing, November 2011<br />

Test of causality, realistic<br />

scenario<br />

Potential interaction between<br />

treatments, activity level of<br />

individuals not controlled for,<br />

individuals cannot be<br />

identified<br />

HD and HB are comparable<br />

and outperform random by<br />

+60% and LD by factor 3<br />

Large real-world network<br />

based on firm data,<br />

identification of determinants<br />

Missing edges between<br />

noncustomers (HB could not<br />

be tested), causality cannot<br />

be tested<br />

HD outperforms random by<br />

factor 2 and LD by factor 8–9<br />

We also varied the payment levels for successful referrals<br />

to account for the potential effects of extrinsic motivation<br />

(incentive for sharing yes/no). When they received<br />

no incentive for sharing, participants earned remuneration<br />

only if they correctly entered the secret token (∼.40 EUR<br />

per token; for detailed instructions, see the Web Appendix<br />

at http://www.marketingpower.com/jmnov11). In contrast,<br />

under the incentives-for-sharing condition, they received<br />

an additional monetary reward when they were named<br />

as a referrer (.25 EUR per correctly entered token and<br />

.20 EUR per referral; for detailed instructions, see the<br />

Web Appendix).<br />

Therefore, we systematically varied the 4×2×2 = 16 different<br />

treatments, with two replications per treatment. The<br />

limitations of the social networking platform’s messaging<br />

system prevented us from replicating four specific treatments,<br />

so we obtained a total of 28 experimental settings<br />

(the potential maximum was 4 × 2 × 2 × 2 = 32 experimental<br />

settings). Although we systematically varied the treatments,<br />

we placed the low-incentive-for-sharing before the<br />

high-incentive-for-sharing settings to avoid confusing participants<br />

with different incentive instructions. We always


seeded one token and then captured all responses two<br />

weeks after the seeding.<br />

Overall, 55% of the participants actively spread or<br />

entered unique tokens, resulting in 1155 responses. The<br />

average number of tokens spread per experimental setting<br />

was 41.25, with a standard deviation of 19.21. To<br />

compare the success of the strategies, we use a randomeffects<br />

logistic regression analysis that accounts for individual<br />

behavioral differences according to each participant’s<br />

responsiveness in each experimental setting. We use the<br />

number of correctly entered tokens as a dependent variable,<br />

which can be 1 if the token was correctly reported and 0 if<br />

otherwise. With 120 participants and 28 experimental settings,<br />

we obtained 3360 observations. As the independent<br />

variables, we included dummy-coded treatment variables<br />

that reflect our full-factorial design, as we detail in Table 3.<br />

The model achieves a pseudo R-square of 15.5%.<br />

The proportion of unexplained variance accounted for by<br />

subject-specific differences due to unobserved influences,<br />

labeled �, is greater than 90%. Compared with random<br />

seeding, the high-degree seeding strategy yields a much<br />

higher likelihood of response (odds ratio = 1�53) that is similar<br />

to the high-betweenness seeding strategy (odds ratio =<br />

1�39). In contrast, the low-degree seeding strategy dramatically<br />

decreases the likelihood of response (odds ratio = �19).<br />

Our treatment variable, high seeding (dummy coded<br />

as 0 = 12 seeds and 1 = 24 seeds), positively influences<br />

response likelihood. Furthermore, the type of incentive<br />

offered drives the high odds ratio estimate, which might<br />

explain why extrinsic motivation in the form of monetary<br />

incentives is popular for viral marketing (e.g., recruit-afriend<br />

campaigns offering rewards such as price discounts<br />

or coupons for successful referrers; Biyalogorsky, Gerstner,<br />

and Libai 2001). Finally, the participants who received the<br />

token from the experimenter (“seeded by experimenter”)<br />

exhibited a higher response likelihood, which is not surprising<br />

because the information probability in this case equals 1.<br />

TABLE 3<br />

Individual Probability to Respond (i.e., Entering<br />

the Correct Token at Experimental Website<br />

[Random Effects Logit Model, Study 1])<br />

Variable Odds Ratio SE<br />

Seeding Strategy<br />

Low degree �19∗∗∗ �04<br />

High betweenness 1�39∗ �28<br />

High degree 1�53∗∗ �31<br />

High seeding 1�89∗∗∗ �28<br />

High incentives 38�11∗∗∗ 26�07<br />

Seeded by experimenter 14�36∗∗∗ Random coefficient: User ID<br />

3�74<br />

ln(� 2 u<br />

� 3�36 �17<br />

� u 5�36 �46<br />

P �90 �02<br />

R 2 (pseudo) .16<br />

N 3360<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Notes: Two-tailed significance levels. n�s� = not significant. Reference<br />

category: “random” seeding strategy, “low seeding,”“no<br />

incentives,” and “was not seeded by experimenter.”<br />

To compare the various seeding strategies directly, we<br />

also varied the contrast specifications but left the rest of<br />

the model unchanged, which produced the conditional odds<br />

ratio matrix in Table 4. As Table 4 indicates, both highdegree<br />

and high-betweenness seeding increase response<br />

likelihood, in contrast with the random seeding strategy, by<br />

39%–53%. Compared with the low-degree seeding strategy<br />

(second column of Table 4), all other strategies are five<br />

to eight times more successful. However, a comparison of<br />

the two most successful seeding strategies, high betweenness<br />

and high degree, does not yield significant differences.<br />

This result has key implications for marketing practice, in<br />

that degree centrality as a local measure is much easier to<br />

compute than betweenness centrality, which requires information<br />

about the structure of the entire network.<br />

In summary, we find that the low-degree seeding strategy<br />

is inferior to the other three seeding strategies and that both<br />

high-betweenness and high-degree seeding outperform the<br />

random seeding strategy but yield comparable results. However,<br />

we also acknowledge that this experiment might suffer<br />

from sequential effects, which would limit the validity<br />

of our separate analysis of each experimental setting. The<br />

behavior of a respondent in one experimental setting might<br />

be influenced by his or her experience in prior experimental<br />

settings. This problem is driven by the limited number<br />

of participants in our experiment. Therefore, in Study 2 we<br />

include more participants and avoid sequential effects by<br />

implementing the four seeding strategies simultaneously.<br />

Study 2: Comparison of seeding strategies in a field setting.<br />

In a second field experiment, we focused on the entire<br />

online social network of all students enrolled in the MBA<br />

program at the same university as in Study 1. Thus, the<br />

network boundary is defined by participation in the program.<br />

We collected contact information for 1380 students<br />

(1380 nodes, 4052 edges) by crawling the same social networking<br />

platform to collect information on friendships, and<br />

we then calculated the sociometric measures as in Study 1.<br />

TABLE 4<br />

Conditional Odds Ratios of Seeding Strategies<br />

(Study 1)<br />

Low High High<br />

Degree Random Betweenness Degree<br />

Low degree — �19 ∗∗∗ �13 ∗∗∗ �12 ∗∗∗<br />

Random 5�37 ∗∗∗ — �72 ∗ �65 ∗∗<br />

High<br />

betweenness 7�47 ∗∗∗ 1�39 ∗ — �91 n�s�<br />

High degree 8�19 ∗∗∗ 1�53 ∗∗ 1�10 n�s� —<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Notes: Two-tailed significance levels. n�s� = not significant. Read the<br />

second column as follows: The odds that a person reacts<br />

to the strategy of random seeding is 5.37 times as large as<br />

that for low-degree seeding, 7.47 times as large in the strategy<br />

of high-betweenness seeding as for low-degree seeding,<br />

and 8.19 times as large in the strategy of high-degree seeding<br />

as for low-degree seeding. The conditional odds ratio<br />

of the two seeding strategies relate inversely. For example,<br />

the odds ratios of random and low degree relate as follows:<br />

�19 = 1/5�37.<br />

Seeding Strategies for Viral Marketing / 61


The mean degree centrality (standard deviation) is 5.872<br />

(7.318). Study 2 also reveals a high and significant correlation<br />

between degree centrality in the bounded network<br />

(1380 MBA students at the university) and degree centrality<br />

in the entire network of the social networking platform<br />

(6.2 million unique users in November 2009). The Pearson<br />

correlation of .824 (p < .001) thus suggests that the number<br />

of friends reported is also a good indicator of degree<br />

centrality in abounded network.<br />

As a proxy for the level of activity, we also used the<br />

time since the last profile update in Study 2. We acquired<br />

information about 849 update time stamps (we could not<br />

access 531 due to the privacy restrictions set by users). On<br />

average, users updated their profile 25.7 weeks ago (Mdn =<br />

15�0), and we observed a weak but significant correlation<br />

between degree centrality and time (in weeks) since the last<br />

profile update (r = −�192, p < �01). We also observed a correlation<br />

between betweenness centrality and time since the<br />

last profile update (r = −�154, p < �01). These negative correlation<br />

simply that participants who updated their profiles<br />

more recently (and probably update them more frequently)<br />

are also more central in the social network. In other words,<br />

activity correlates with centrality and may be an additional<br />

determinant of the viral spread of information in this setting.<br />

In terms of gender (805 male, 569 female, and 6 missing<br />

observations), male participants were more central, such<br />

that the average female participant had .92 fewer connections<br />

than the average male (p < �05). However, this gender<br />

difference becomes insignificant if we control for activity.<br />

The experimental setup for Study 2 was somewhat different.<br />

First, the four treatment groups (hubs, bridges, fringes,<br />

or random sample) were all seeded on the same day. Second,<br />

we eliminated the incentive variation, such that we<br />

did not use extrinsic monetary incentives to stimulate participation.<br />

Third, we did not vary the seeding size and<br />

sent a reminder out to the initial seeds seven days after<br />

the initial seeding. The seeding included 95 participants in<br />

each of the four treatments (70 on Day 1, 25 on Day 2),<br />

or 7% of the total network (which is in line with Jain,<br />

Mahajan, and Muller 1995). The seed message contained a<br />

60<br />

40<br />

20<br />

0<br />

FIGURE 1<br />

Development of the Number of Unique Visits Over Time (Study 2)<br />

Entry Page: Cumulative Number of Unique Visits<br />

80<br />

0 2 4 6 8 10 12 14 16 18 20<br />

Time (Days)<br />

62 / Journal of Marketing, November 2011<br />

High degree<br />

High betweenness<br />

Random<br />

Low degree<br />

65<br />

63<br />

40<br />

22<br />

unique URL for a website with a funny video that we produced<br />

about the participants’ university (the landing page<br />

and video were identical for all treatments). By producing<br />

a new video specifically for this second field experiment,<br />

we ensured that the viral marketing stimulus (i.e., content)<br />

was unknown to all participants. Furthermore, we predicted<br />

that the link to the video would be distributed preferentially<br />

to fellow students (from which we obtained mutual<br />

online social network relationships), rather than to others<br />

outside the university’s social network. In other words, the<br />

social network for Study 2 should represent a coherent,<br />

self-contained social community. One MBA student served<br />

as the initiator who seeded the message to others, according<br />

to the chosen seeding strategy. In addition to the link<br />

to the particular entry page, the message indicated that the<br />

addressees could find a funny video about the university<br />

that had just been created by the initiator.<br />

We tracked website visits for the entry pages and video<br />

download pages of the four sites (one for each strategy)<br />

for 19 days. Figure 1 compares the success of the seeding<br />

strategies. The rank order with respect to their success,<br />

across both dependent variables, is consistent with<br />

the results from our first experiment. That is, high-degree<br />

and high-betweenness seeding clearly outperform both lowdegree<br />

and random seeding. For example, in terms of<br />

videos watched, the high-degree seeding strategy yielded<br />

more than twice the number of responses than did random<br />

seeding. Information about social position thus made<br />

it possible to more than double the number of responses.<br />

We also estimated two random-effects linear models (one<br />

for the entry page, one for the video page) in which we<br />

treated each of the 19 days as a unit of observation, for<br />

which we have four observations. The dependent variable<br />

is thus the number of unique visits for the entry page and<br />

the number of unique video requests from the video page.<br />

We included the seeding and reseeding (reminder) days as<br />

dummy variables and added another dummy variable to<br />

account for weekends. The seeding strategies also are coded<br />

as dummy variables, and the experimental day is a unit<br />

specific random coefficient. Table 5 illustrates the results.<br />

Video Page: Cumulative Number of Unique Visits<br />

80<br />

60<br />

40<br />

20<br />

Seeding activity<br />

High degree<br />

43<br />

35<br />

High betweenness<br />

Random 17<br />

12<br />

Low degree<br />

0<br />

0 2 4 6 8 10 12 14 16 18 20<br />

Time (Days)


TABLE 5<br />

Number of Visits per Day (Random Effects Model, Study 2)<br />

Entry Page Video Page<br />

Unique Visits Unique Visits<br />

Variable Coefficient SE Coefficient SE<br />

High-degree seeding 2�263 ∗∗∗ �766 1�623 ∗∗∗ �547<br />

High-betweenness seeding 2�158 ∗∗∗ �766 1�211 ∗∗ �547<br />

Random seeding �947 n�s� �766 �263 n�s� �547<br />

Seeding day or reseeding 7�128 ∗∗∗ 1�276 4�005 ∗∗∗ �727<br />

Weekend −1�026 n�s� 1�569 −�636 n�s� �794<br />

Intercept −�127 n�s� �911 −�078 n�s� �524<br />

Random Coefficient: Experimental Day<br />

� e 2.375 1.642<br />

P .519 .290<br />

R 2 (overall) .475 .436<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Notes: Two-tailed significance levels. n�s� = not significant. Reference categories: “low-degree seeding” and “weekdays” and “no seeding day.”<br />

The models for both the entry and video pages are highly<br />

significant, with explained overall variances (adjusted R 2 )<br />

of 47.5% and 43.6%. The results in Figure 1 confirm our<br />

previous observations: High-degree and high-betweenness<br />

seeding yield comparable results and are three times more<br />

successful than low-degree seeding and 60% more successful<br />

than random seeding. Days with seeding or reseeding<br />

activities yield more unique visits. Responsiveness declined<br />

on weekends (albeit insignificantly), perhaps due to the<br />

overall higher level of online activity by these students on<br />

weekdays.<br />

In summary, Study 2 supports our findings from Study 1<br />

that seeding to hubs and bridges is preferable to seeding<br />

to fringes. However, we also note the potential for interactions<br />

among the activities associated with the four seeding<br />

strategies in Study 2. For example, a participant might have<br />

watched the video after receiving a message from seeding<br />

strategy A and then receive a nearly identical message from<br />

seeding strategy B, in which case this participant is unlikely<br />

to click the link again to watch the same video. Thus,<br />

seeding strategies that foster faster diffusion may have an<br />

advantage that could bias the result and lead to over estimations<br />

of the success of high-degree and high-betweenness<br />

seeding in contrast with random and low-degree seeding.<br />

However, in Study 1, such crossings were not possible as a<br />

result of the sequential timing, and the results remained the<br />

same. Neither Study 1 nor Study 2 identifies the reasons for<br />

the superiority of specific seeding strategies. In addition,<br />

we cannot distinguish between first- and second-generation<br />

referrals. We address these shortcomings in Study 3.<br />

Comparison of the Effect of Seeding Strategies<br />

on the Determinants of Social Contagion in a<br />

Real-Life Viral Marketing Campaign (Study 3)<br />

For Study 3, a mobile phone service provider stimulated<br />

referrals (through text messages) to attract new customers.<br />

The provider tracked all referrals, so we can compare the<br />

economic success of different seeding strategies and analyze<br />

the influence of the corresponding sociometric measures<br />

on all determinants of social contagion (Table 1) in<br />

a real-life setting. This helps us identify the reasons for<br />

any differences. Thus, Study 3 enables us to decompose<br />

the effect of the different determinants that drive the social<br />

contagion process, including participation probability P i,<br />

the used reach n i, the mean conversion rate w i of all referrals<br />

made by ion the expected number of referrals R i, and<br />

the expected number of successful referrals SR i. The viral<br />

marketing campaign of the mobile phone service provider<br />

featured text messages sent to the entire customer base<br />

(n = 208�829 customers), promising a 50% higher reward<br />

than the regular bonus of E 10 worth of airtime for each<br />

new customer referred in the next month. In total, 4549 customers<br />

participated in the campaign, initiating 6392 firstgeneration<br />

referrals, which was a 50% increase over the<br />

average number of referrals. We anticipate that social contagion<br />

works through belief updating, as prospective customers<br />

talk to adopters about the product. Furthermore, in<br />

Becker’s (1970) terms, we classify this product as a lowrisk<br />

offering (cf. trials of untested drugs).<br />

Our analysis of the social contagion process is based on a<br />

rich data set; each referral activity was logged in the online<br />

referral system of the company, because customers had to<br />

initiate the referral messages to friends online. Successful<br />

referrals were confirmed during the registration process of<br />

the new customers, who had to identify their referrer to trigger<br />

the payment of the referral premium. Thus, we gathered<br />

information about whether customers acted on the stimulus<br />

of the referral, captured by the variable program participation<br />

P i, as well as the number of referrals R i and the<br />

number of successful referrals SR i. The mean conversion<br />

rate per referrer w i can be inferred from a comparison of<br />

R i and SR i. We used individual-level communication data<br />

and the number of text messages to others to calculate the<br />

(external) degree centrality. (In total, we evaluated more<br />

than 100 million connections. 1 ) We assumed that any telephone<br />

call or text message between people (independent of<br />

the direction) reflected social ties. Thus, degree centrality<br />

1 All individual-level data were made anonymous with a multistage<br />

encryption process, undertaken by the firm before the analysis.<br />

At no point was any sensitive customer information, such as<br />

names or telephone numbers, disclosed.<br />

Seeding Strategies for Viral Marketing / 63


equals a count of the total number of unique communication<br />

relationships. However, the service can only be referred<br />

to current non customers, so the degree centrality metric<br />

accounts only for ties that customers had to people outside<br />

the service network at the beginning of the viral marketing<br />

campaign, which makes it a form of external degree<br />

centrality. We lacked information about the relationships<br />

of people who were not customers, so we could not measure<br />

betweenness centrality and test the high-betweenness<br />

seeding strategy in Study 3.<br />

We used the following customer characteristics as covariates:<br />

demographic information including age (in years) and<br />

gender (1 = female, and 0 = male); service-specific characteristics,<br />

such as customer tenure (i.e., length of the relationship<br />

with the company in months); and the tariff plan.<br />

We operationalized the tariff plan with a dichotomous variable<br />

to indicate whether the customer chose a community<br />

tariff (=1, including a reduced per-minute price for<br />

calls within the network) or a one-price tariff (=0). Furthermore,<br />

we used two measures of customers’ trust in the<br />

service: payment type (dichotomous variable: automatic =<br />

1, and manual = 0) and refill policy (dichotomous variable:<br />

automatic = 1, and manual = 0). In the case of automatic<br />

payment and refill, customers provided credit card details to<br />

the service provider. Finally, we included information about<br />

the acquisition channel for each customer (1 = offline/retail,<br />

and 0 = online). As additional controls, we include information<br />

on the individual service usage of the customer,<br />

namely, average monthly airtime (in minutes) and monthly<br />

short message service (SMS)—that is, the average monthly<br />

number of SMS sent by a customer.<br />

Our model reflects the two-stage process for each participant,<br />

who first decides whether to participate (P i) and<br />

then chooses to what extent to participate (n i�. A specific<br />

characteristic of the first stage is the relatively large share<br />

of zeros (i.e., nonparticipants), whereas observed values for<br />

the second stage are count measures and highly skewed.<br />

This data structure requires specific two-stage regression<br />

models: either inflation models, such as the zero-inflated<br />

Poisson (ZIP) regression (Lambert 1992), or hurdle models,<br />

such as the Poisson-logit hurdle regression (PLHR) model<br />

(Mullahy 1986). We use a PLHR model, which combines<br />

a logit model to account for the participation decision and<br />

a zero-truncated Poisson regression to analyze the actual<br />

outcomes of participation (e.g., number of successful referrals).<br />

2 In our PLHR specification, the binary variable P i<br />

indicates whether individual i participates in the referral<br />

program (hurdle or logit model). In addition, Used Reach n i<br />

indicates how many referrals the individual i initiates, conditional<br />

on the decision to participate (P i = 1). As an extension,<br />

Converted Reach CR = �n i × w i� indicates how many<br />

successful referrals individual i initiates, again conditional<br />

on the decision to participate. Note that Used Reach n i<br />

and Converted Reach CR i are equivalent to Referrals R i<br />

and Successful Referrals SR i, respectively, conditional on<br />

a program participation probability P i = 1. These variables<br />

2 We choose PLHR over ZIP because the logit stage of the former<br />

is designed to determine what leads to participation (i.e., identifying<br />

referrers, in which we are interested), whereas the inflation<br />

stage of ZIP tries to detect “sure zeros” (i.e., nonparticipants).<br />

64 / Journal of Marketing, November 2011<br />

provide the dependent variables in the Poisson regression<br />

of our PLHR specification.<br />

Let P∗ i be the latent variable related to Pi, n∗ i be the<br />

censored variable related to ni, and CR∗ = �n∗ i × wi�∗ be<br />

the censored variable related to CR = �n∗ i × wi�. Together<br />

with the explanatory variable of (external) degree centrality<br />

and the covariates (age, gender, payment type, refill policy,<br />

acquisition channel, and customer tenure), the PLHR can<br />

be specified as follows:<br />

�<br />

1 if P<br />

Pi =<br />

∗ i > 0<br />

where P∗<br />

i<br />

0 otherwise� = � (3)<br />

P0i + �Pij × XPij + �Pi� (4)<br />

and<br />

(5)<br />

n i =<br />

�<br />

n ∗ i<br />

if P ∗ i >0<br />

0 otherwise�<br />

CR = �n i × w i� =<br />

where n∗<br />

i = � UR0i +� URij ×X URij +� URi�<br />

�<br />

�ni × wi�∗ if CR∗ i > 0<br />

0 otherwise�<br />

where �n i × w i� ∗ = � CR0i + � CRij × X CRij + � CRi�<br />

Thus, X ij contains the explanatory variables j (i.e., degree<br />

centrality and the covariates) and the error terms � Pi, � URi,<br />

and � CRi, which represent unobserved influences on participation<br />

probability, used reach, and the number of successful<br />

referrals.<br />

Seeding strategies in first-generation models. In a first<br />

step, we restricted our analysis to first-generation models;<br />

we only considered referrals directly initiated by customers<br />

who received the seeding stimulus during the viral marketing<br />

campaign. Table 6 contains the parameter estimates<br />

of the PHLR model. The results can be interpreted in two<br />

stages: first, what drives the participation of seeded customers<br />

in the viral marketing campaign (logit component =<br />

LC) and, second, among these participants, what influences<br />

the number of referrals and successful referrals (Poisson<br />

component = PC). With regard to the covariates’ impact<br />

on program participation, we find significant effects of the<br />

demographic variables gender (�LC 2n∗ = −�2171, p < �01) and<br />

age (�LC 3n∗ = −�0209, p < �01), indicating that male and older<br />

customers are more likely to participate, as are customers<br />

with short customer tenures who have just recently adopted<br />

the service (�LC 8n∗ = −�0016, p < �01). The latter finding aligns<br />

with cognitive dissonance theory, in that these customers<br />

might communicate shortly after their purchase decision<br />

to reduce dissonance (Festinger 1957). Furthermore, we<br />

found that customers acquired online are more strongly<br />

engaged in the (online-based) referral program than customers<br />

acquired through the retail channel (�LC 6n∗ = −�9843,<br />

p < �01). A one-price tariff seems easier to communicate;<br />

seeded customers with that tariff option are more likely to<br />

participate in the referral program (�LC 7n∗ = −�1433� p < �01).<br />

With regard to the usage covariates, we find positive and<br />

significant values for monthly airtime (�LC 9n∗ = �0007� p < �01)<br />

and monthly SMS (�LC 10n∗ = �0010� p < �01). The influences<br />

of most covariates are comparable between the logit and<br />

Poisson regression stages, except the acquisition channel,<br />

in that retail customers are less likely to participate in the<br />

program, but if they do, they exhibit significantly greater


activity than online customers (�PC 6n∗ = �7388� p < �01). The<br />

same applies to the usage covariates: Here, monthly air-<br />

time (�PC 9n∗ = −�0007� p < �1) and monthly SMS (�PC 10n∗ =<br />

−�0018� p < �01) show negative effects on participation.<br />

However, the influence of (external) degree centrality<br />

varies between the stages of the model. In the logit regression<br />

stage, degree centrality has a positive and significant<br />

influence on the likelihood to participate P i in the refer-<br />

ral program (�LC 1n∗ = �0032� p < �01). Confirming the results<br />

of Studies 1 and 2, this finding shows that customers with<br />

high degree centrality are more likely to participate than<br />

those with low degree centrality (average degree centrality<br />

of participants = 45�3 vs. nonparticipants = 36�5). However,<br />

in the Poisson regression stage that analyzes only<br />

the group of active referrers, the effect of degree central-<br />

ity is mixed. We find a significant, positive effect on used<br />

reach (�PC 1n∗ = �0012� p < �01), such that customers with highdegree<br />

centrality are not only more likely to participate but<br />

also more active when participating in the viral marketing<br />

campaign. However, we find no significant effect of degree<br />

TABLE 6<br />

Determinants of Number of Referrals, Number of Successful Referrals, and Influence Domain<br />

(Poisson-Logit Hurdle Regression Models, Study 3)<br />

Converted Reach Conditional Influence<br />

Used Reach n ∗ CR = �n ∗ w� ∗ Domain ID T<br />

i<br />

Variable Coefficient SE Coefficient SE Coefficient SE<br />

Logit Component<br />

Degree centrality � 1 �0022 ∗∗∗ �0003 �0021 ∗∗∗ �0004 �0021 ∗∗∗ �0004<br />

Covariates<br />

Gender � 2 −�2287 ∗∗∗ �0318 −�2527 ∗∗∗ �0337 −�2527 ∗∗∗ �0337<br />

Age � 3 −�0202 ∗∗∗ �0013 −�0190 ∗∗∗ �0014 −�0189 ∗∗∗ �0014<br />

Payment type � 4 �0913 ∗∗ �0389 �0668 n�s� �0412 �0668 n�s� �0412<br />

Refill policy � 5 −�0122 n�s� �0376 �0174 n�s� �0396 �0174 n�s� �0396<br />

Acquisition channel � 6 −�9848 ∗∗∗ �0506 −1�0535 ∗∗∗ �0545 −1�0530 ∗∗∗ �0545<br />

Tariff plan � 7 −�1371 ∗∗∗ �0395 −�1253 ∗∗∗ �0419 −�1253 ∗∗∗ �0420<br />

Customer tenure � 8 −�0016 ∗∗∗ �0001 −�0016 ∗∗∗ �0001 −�0016 ∗∗∗ �0001<br />

Monthly airtime � 9 �0007 ∗∗∗ �0002 �0007 ∗∗∗ �0002 �0007 ∗∗∗ �0002<br />

Monthly SMS � 10 �0010 ∗∗∗ �0002 �0009 ∗∗∗ �0002 �0010 ∗∗∗ �0002<br />

Intercept −2�3825 ∗∗∗ �0727 −2�534 ∗∗∗ �0770 −2�5341 ∗∗∗ �0770<br />

Poisson Component<br />

Degree centrality � 1 �0026 ∗∗∗ �0006 �0001 n�s� �0033 −�0025 ∗∗∗ �0007<br />

Covariates<br />

Gender � 2 −�1539 ∗∗∗ �0519 −�1177 n�s� �2206 −�3323 ∗∗∗ �0496<br />

Age � 3 −�0098 ∗∗∗ �0020 −�0104 n�s� �0087 −�0112 ∗∗∗ �0019<br />

Payment type � 4 −�3908 ∗∗∗ �0581 −�1807 n�s� �2492 −�1379 ∗∗ �0544<br />

Refill policy � 5 −�3417 ∗∗∗ �0761 −�1369 n�s� �2819 �0033 n�s� �0608<br />

Acquisition channel � 6 �7408 ∗∗∗ �0561 �5902 ∗∗ �2592 �6273 ∗∗∗ �0548<br />

Tariff plan � 7 −1�0740 ∗∗∗ �0473 −1�1426 ∗∗∗ �2067 −�5322 ∗∗∗ �0466<br />

Customer tenure � 8 −�0007 ∗∗∗ �0002 −�0007 n�s� �0007 −�0013 ∗∗∗ �0001<br />

Monthly airtime � 9 −�0007 ∗ �0004 −�0000 n�s� �0000 �0000 n�s� �0003<br />

Monthly SMS � 10 −�0018 ∗∗∗ �0005 −�0001 n�s� �0019 �0005 n�s� �0004<br />

Intercept �9811 ∗∗∗ �0941 −1�4633 ∗∗∗ �4181 1�1008 ∗∗∗ �0905<br />

Log-likelihood value −25�163 −19�850 −23�723<br />

BIC 50,596 39,969 47,714<br />

N 208,829<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Notes: Two-tailed significance levels. n�s� = not significant.<br />

centrality on the referral success of active referrers (� PC<br />

1CR =<br />

−�0002� n�s�).<br />

This result is further confirmed when we analyze the<br />

mean conversion rate of referrals per referrer w i = CR i/n ∗ i<br />

(see Table 7). Again, we do not find a significant effect<br />

of degree centrality for active referrers (� 1w = �0001� n�s�).<br />

Thus, our results offer no support for the assumption that<br />

participating central customers are more persuasive referrers<br />

or better selectors of potential referral targets.<br />

Next, considering that viral marketing campaigns can<br />

be costly, we attempt to identify customers who are most<br />

likely to participate and generate (successful) referrals. We<br />

use the estimated participation probability calculated from<br />

the results of the selection model (see Table 6, “Logit Component”)<br />

to group the full customer base into cohorts and<br />

then compare these cohorts according to their observed participation,<br />

referral, and conversion rates and degree centrality<br />

(see Table 8). The top 5000 cohort corresponds to<br />

a high-degree and the bottom 5000 cohort to a low-degree<br />

seeding strategy; the results in the “Average” column correspond<br />

to random seeding.<br />

Seeding Strategies for Viral Marketing / 65


TABLE 7<br />

Determinants of Conversion Rates, Active<br />

Referrers (Poisson Regression, Study 3)<br />

Poisson Regression Model<br />

Conversion w = CR/n ∗<br />

Variable Coefficient SE<br />

Degree centrality � 1 −�0001 n�s� �0004<br />

Covariates<br />

Gender � 2 �0788 ∗∗ �0350<br />

Age � 3 �0047 ∗∗∗ �0015<br />

Payment type � 4 �1104 ∗∗∗ �0431<br />

Refill policy � 5 �1079 ∗∗∗ �0409<br />

Acquisition channel � 6 −�4951 ∗∗∗ �0616<br />

Tariff plan � 7 �4638 ∗∗∗ �0472<br />

Customer tenure � 8 �0002 ∗ �0001<br />

Intercept −1�2014 ∗∗∗ �0848<br />

Log-likelihood value −5� 074<br />

N 4,549<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Note: Two-tailed significance levels. n�s� = not significant. For the<br />

Poisson regression model, we used ln�n� as offset variable.<br />

The results in Table 8 clearly confirm the positive<br />

correlation between degree centrality and the success of<br />

viral marketing: As the estimated participation probability<br />

increases, observed participation, referral, and conversion<br />

rates (i.e., total number of participants, referrals, or successful<br />

referrals divided by number of seeded customers in<br />

the cohort) and degree centrality increase as well. Thus,<br />

the participation rate of the top 5000 cohort is a multiple<br />

of that of the bottom 5000 cohort (4.4% vs. .5%), with a<br />

much higher average degree centrality (70.8 vs. 18.0). A<br />

high-degree seeding strategy would be nearly nine times<br />

as successful as a low-degree strategy. Compared with the<br />

average value of a random strategy, the top 5000 cohort participation<br />

and degree centrality are twice as high; therefore,<br />

targeting hubs doubles the performance of random seeding<br />

for a sample of the same size. Thus, Study 3 clearly shows<br />

the significant, positive effect of degree centrality on viral<br />

marketing participation and activity, in strong support of a<br />

high-degree seeding strategy. However, the results of the<br />

Poisson regression model do not indicate higher referral<br />

success of hubs within the group of active referrers.<br />

Seeding strategies in multiple-generation models. In the<br />

second step of our empirical analysis, we extended the measure<br />

of success to account for a fuller range of the effects<br />

of seeding efforts by including more than one generation<br />

of referrals. First-generation referrals initiate a viral process<br />

that should continue in further generations. The extent<br />

of this viral branching may differ across seeding strategies,<br />

due to their ability to reach different parts of the social network.<br />

Thus, the optimal seeding strategy might change if<br />

we consider multiple generations.<br />

To capture this form of success, we measured all subsequent<br />

referrals that originated from a first-generation referral<br />

during the campaign. We limited the observation period<br />

to 12 months because the company repeated the referral<br />

campaign 13 months after the initial seeding. During our<br />

observation period, the company did not engage in other<br />

promotions that directly focused on referrals, nor did we<br />

find any anomalies (e.g., drastic increases or decreases) in<br />

company-owned or competitive marketing spending.<br />

In the first year after the campaign, 20.8% of all firstgeneration<br />

referrals became active referrers themselves, and<br />

5.8% did so multiple times. We observed viral referral<br />

chains with a maximum length of 29 generations; on average,<br />

every first-generation referral during the campaign led<br />

to .48 additional referrals.<br />

The dependent variable for this analysis is the influence<br />

domain of all successful referrals of a specific<br />

first-generation customer, which equals the number of<br />

successful first-generation referrals, plus the number of successful<br />

referrals in successive generations during the subsequent<br />

12 months. 3 For example, Figure 2 depicts the<br />

influence domain of a referral customer X that spans 22<br />

additional successful referrals over seven generations.<br />

3 Note that, by definition, there is no overlap of influence<br />

domains between two origins; every referred customer has an indegree<br />

of 1, and only one specific referrer is rewarded for every<br />

new customer.<br />

TABLE 8<br />

Relationship of Conversion Rates and Degree Centrality, Full Sample (Study 3)<br />

Customer Cohort<br />

(According to Estimated Participation Probabilities)<br />

Top 5000 Top 10,000 Top 20,000 Top 50,000 Bottom 5000 Average<br />

Participants<br />

Total participation 220 378 671 1385 24 —<br />

Participation rate 4�4% 3�8% 3�4% 2�8% �5% 2�2%<br />

Referrals<br />

Total referrals 292 489 856 1783 26 —<br />

Referral rate 5�8% 4�9% 4�3% 3�6% �5% 3�0%<br />

Successful referrals<br />

Total conversions 191 330 598 1233 19 —<br />

Conversion rate 3�8% 3�3% 3�0% 2�5% �4% 2�0%<br />

Average degree centrality 70�83 60�21 52�13 45�42 18�01 36�48<br />

Notes: Top (Bottom) 5,000/10,000/� � � refer to the cohort of customers with the highest (lowest) estimated participation probabilities, according<br />

to the coefficient estimates of the logit component reported in Table 6. We calculated rates by dividing the total number of participants,<br />

referrals, or successful referrals by the total number of customers in the cohort.<br />

66 / Journal of Marketing, November 2011


3<br />

FIGURE 2<br />

Influence Domain of a Referral Campaign<br />

Participant (Study 3)<br />

2<br />

1<br />

1<br />

Customer X<br />

Initial campaign (origin)<br />

stimulus<br />

2<br />

2<br />

3 3<br />

3<br />

3<br />

4<br />

3 3 4<br />

4<br />

5<br />

7<br />

6<br />

6<br />

7<br />

Referral<br />

generations<br />

7<br />

Customer Y<br />

The parameter estimates of the PLHR model for this<br />

multiple-generation model appear in the right-hand column<br />

of Table 6. The dependent variable IDT i is conditional<br />

on program participation (Pi = 1). When we compare the<br />

regression model parameters across the different dependent<br />

variables, we find similar results for Influence Domain<br />

IDT i and Used Reach ni (e.g., no significant effect of the<br />

usage covariates) but with one important difference: Our<br />

focal variable, degree centrality, is negative (�PC 1ID = −�00205,<br />

p < .01) in the Poisson regression model. That is, among the<br />

participants, more central customers have a smaller influence<br />

domain. The observed network structure of the referral<br />

processes offers a potential explanation of this surprising<br />

TABLE 9<br />

Determinants of Unconditional Influence Domain (OLS Model, Study 3)<br />

OLS Model Unconditional<br />

Poisson Regression<br />

Model Unconditional<br />

Influence Domain (ID R<br />

i ) Influence Domain �IDR<br />

Variable Standard Coefficient SE Coefficient SE<br />

Degree centrality � 1 �010 ∗∗∗ �000 �002 ∗∗∗ �000<br />

Covariates<br />

Gender � 2 −�015 ∗∗∗ �002 −�358 ∗∗∗ �028<br />

Age � 3 −�024 ∗∗∗ �000 −�024 ∗∗∗ �001<br />

Payment type � 4 �001 n�s� �002 �013 n�s� �033<br />

Refill policy � 5 �000 n�s� �002 �003 n�s� �033<br />

Acquisition channel � 6 −�025 ∗∗∗ �002 −�680 ∗∗∗ �039<br />

Tariff plan � 7 −�016 ∗∗∗ �002 −�433 ∗∗∗ �030<br />

Customer tenure � 8 −�031 ∗∗∗ �004 −�002 ∗∗∗ �000<br />

Intercept �091 ∗∗∗ �004 −1�509 ∗∗∗ �058<br />

R 2 (pseudo) .05 .03<br />

N 208,829 208,829<br />

∗ p < �1.<br />

∗∗ p < �05.<br />

∗∗∗ p < �01.<br />

Notes: Two-tailed significance levels. n�s� = not significant.<br />

result. For hubs, we mostly observe short referral chains (if<br />

at all), whereas fringe customers who participate in the viral<br />

marketing campaign demonstrate significantly longer referral<br />

chains. For example, in Figure 2, the fringe customer X<br />

reacts to the campaign and refers the service. Within two<br />

generations, this referral reaches actor Y, who initiates a<br />

total of 15 additional referrals, which increases the influence<br />

domain of X to 22.<br />

Because we find a positive effect of high degree centrality<br />

in the selection model but a negative effect in the<br />

regression model, the overall effect of degree centrality<br />

remains unclear. For a simple test, we performed both an<br />

ordinary least squares (OLS) and a simple Poisson regres-<br />

sion, with the unconditional Influence Domain ID R i<br />

as the<br />

dependent variable for the complete sample of 208,829 customers<br />

who received the viral marketing campaign stimulus.<br />

According to the results in Table 9, the standardized<br />

beta for degree centrality is positive and significant in both<br />

regressions (�OLS 1ID = �010� �PR<br />

1ID = �002� p < �001), such that the<br />

overall effect of high degree centrality as a selection criterion<br />

for seeding a viral marketing campaign is positive.<br />

Therefore, high-degree seeding remains the more successful<br />

strategy, even if we account for a multiple-generation<br />

viral process.<br />

Robustness checks. To check the robustness of our findings,<br />

we also analyzed our data with a set of alternative<br />

approaches, including a Poisson-based (ZIP regression) and<br />

probit–OLS combinations (e.g., Tobit Type II models). Our<br />

core results pertaining to the influence of degree centrality,<br />

including its positive influence on the selection stage<br />

to determine the likelihood of participation and its negative<br />

effect on the influence domain in the regression stage, hold<br />

across all tested models. The results also remain unchanged<br />

when we incorporate alternative individual-level covariates,<br />

such as monthly mobile charges, that represent the attractiveness<br />

of a customer to the provider.<br />

Unlike Studies 1 and 2, Study 3 does not allow us to<br />

assess the causal effect of the seeding strategy on referral<br />

Seeding Strategies for Viral Marketing / 67<br />

i �


success unambiguously. However, it reflects a real-life marketing<br />

application and is based on detailed firm data, such<br />

that it strikingly illustrates the power of network information<br />

in real-life situations.<br />

Research Contribution<br />

General Discussion<br />

Inspired by conflicting recommendations in previous studies<br />

regarding optimal seeding strategies for viral marketing<br />

campaigns, this article empirically compares the performance<br />

of various proposed strategies, examines the magnitude<br />

of differences, and identifies determinants that are<br />

responsible for the superiority of a particular seeding strategy.<br />

To the best of our knowledge, such an experimental<br />

comparison of seeding strategies is unprecedented; previous<br />

literature is based solely on mathematical models and computer<br />

simulations. Our real-life application provides some<br />

answers to controversies about whether hubs are harder<br />

to convince, whether they make use of their reach, and<br />

whether they are more persuasive.<br />

Marketers can achieve the highest number of referrals,<br />

across various settings, if they seed the message to hubs<br />

(high-degree seeding) or bridges (high-betweenness seeding).<br />

These two strategies yield comparable results and both<br />

outperform the random strategy (+52%) and are up to eight<br />

times more successful than seeding to fringes (low-degree<br />

seeding). The superiority of the high-degree seeding strategy<br />

does not rest on a higher conversion rate due to a higher<br />

persuasiveness of hubs but rather on the increased activity<br />

of hubs, which is in line with previous findings (e.g.,<br />

Iyengar, Van den Bulte, and Valente 2011; Scott 2000).<br />

This finding is persistent even when we control for revenue<br />

of customers and thus demonstrates the importance<br />

of social structure beyond customer revenues and customer<br />

loyalty.<br />

Research that suggests a low-degree seeding strategy is<br />

usually based on the central assumption that highly connected<br />

people are more difficult to influence than less connected<br />

people because highly connected people are subject<br />

to the influence of too many others (e.g., Watts and Dodds<br />

2007). Our results (Studies 1 and 3) reject this assumption<br />

and underline Becker’s (1970) suggestion: Hubs are more<br />

likely to engage because viral marketing works mostly<br />

through awareness caused by information transfer from previous<br />

adopters and through belief updating, especially for<br />

low-risk products. The low perceived risk means hubs do<br />

not hesitate before participating. Furthermore, when social<br />

contagion occurs mostly at the awareness stage, the possible<br />

disproportionate persuasiveness of hubs is irrelevant. As<br />

long as the social contagion occurs at the awareness stage<br />

through simple information transfer, hubs are not more<br />

persuasive than other nodes (Godes and Mayzlin 2009;<br />

Iyengar, Van den Bulte, and Valente 2011).<br />

Our analysis of a mobile service provider’s viral marketing<br />

campaign reveals that hubs make slightly more use of<br />

their reach potential. Furthermore, for the group of participating<br />

customers, we find a negative influence of greater<br />

connectivity on the resulting influence domains. Although<br />

in epidemiology studies infectious diseases spread through<br />

hubs, we find that well-connected people do not use their<br />

68 / Journal of Marketing, November 2011<br />

greater reach potential fully in a marketing setting. Spreading<br />

information is costly in terms of both time invested<br />

and the effort needed to capture peers’ attention. Furthermore,<br />

hubs may be less likely to reach other previously<br />

unaffected central actors, such that they are limited in their<br />

overall influence domain. The compelling findings from<br />

sociology and epidemiology thus appear to have been incorrectly<br />

transferred to targeting strategies in viral marketing<br />

settings.<br />

Nevertheless, the social network remains a crucial determinant<br />

of optimal seeding strategies in practice because<br />

a social structure is much easier to observe and measure<br />

than communication intensity, quality, or frequency. Furthermore,<br />

we find robust results even when we control for<br />

the level of communication activity. Therefore, companies<br />

should use social network information about mutual relationships<br />

to determine their viral marketing strategy.<br />

Managerial Implications<br />

Viral marketing is not necessarily an art rather than a<br />

science; marketers can improve their campaigns by using<br />

sociometric data to seed their viral marketing campaigns.<br />

Our multiple studies show that information about the social<br />

structure is valuable, in that seeding the “right” consumers<br />

yields up to eight times more referrals than seeding the<br />

“wrong” ones. In contrast with random seeding, seeding<br />

hubs and bridges can easily increase the number of successful<br />

referrals by more than half. Thus, it is essential<br />

for marketers to adopt an appropriate seeding strategy and<br />

use sociometric data to increase their profits. We conclude<br />

that adding metrics related to social positions to customer<br />

relationship management databases is likely to improve targeting<br />

models substantially.<br />

Many companies already have implicit information about<br />

social ties that they could use to calculate explicit sociometric<br />

measures. Telecommunication providers can exploit<br />

connection data (as we did in Study 3), banks possess<br />

data about money transfers, e-mail providers might analyze<br />

e-mail exchanges, and companies can evaluate behaviors<br />

in company-owned forums. Many companies also have<br />

indirect access to information on social networks, such as<br />

Microsoft through Skype or Google through its Google<br />

mail services, and they could begin to use information<br />

obtained this way (e.g., Hill, Provost, and Volinsky 2006;<br />

Aral, Muchnik, and Sundararajan 2009). Such network<br />

information is further available in the form of friendship<br />

data obtained from online social communities such as Facebook<br />

or LinkedIn (e.g., Hinz and Spann 2008).<br />

Remarkably, we reveal that to target a particular subnetwork<br />

(e.g., students of a particular university, Study 2) with<br />

a viral marketing message, the use of the respective subnetwork’s<br />

sociometric measures is not absolutely required to<br />

implement the desired seeding strategies. Instead, because<br />

the sociometric measures of subnetworks and their total<br />

network are highly correlated, marketers can use the sociometric<br />

measures of the total network, without undertaking<br />

the complex task of determining exact network boundaries.<br />

Conversely, this appealing result also allows marketers to<br />

feel confident in inferring the connectivity of a person in an<br />

overall network from information about his or her connectivity<br />

in a natural subnetwork. Moreover, because betweenness<br />

centrality requires knowledge about the structure of


the entire network, as well as a complex, time-consuming<br />

computation, degree centrality seems to be the best sociometric<br />

measure for marketing practice (see also Kiss and<br />

Bichler 2008).<br />

According to these insights, marketers should pick highly<br />

connected persons as initial seeds if they hope to generate<br />

awareness or encourage transactions through their viral<br />

marketing campaigns because these hubs promise a wider<br />

spread of the viral message. As long as the social contagion<br />

operates at the awareness stage through information transfer,<br />

we do not observe that hubs are more persuasive. Thus,<br />

the sociometric measure of degree centrality cannot be used<br />

to identify persuasive seeding points. The use of demographics<br />

and product-related characteristics (see Table 7)<br />

seems more promising for this purpose. Study 1 reveals<br />

that monetary incentives for referring strongly increase the<br />

spread of viral marketing messages, which supports the use<br />

of such incentives. However, they would also make viral<br />

marketing more costly than is commonly assumed.<br />

Finally, expertise in the domain of social networks is<br />

valuable for seeding purposes. Thus, online communities<br />

such as Facebook might begin to offer information on<br />

members’ social positions to third-party marketers or provide<br />

the option to seed to a specific target group according<br />

to sociometric measures. Specialized service providers<br />

might adopt a similar idea to tailor their offerings with<br />

respect to optimal seeding. Business models might reflect<br />

social network information collected from communication<br />

relationships or domain expertise in certain subject<br />

domains, such that they target the most highly connected<br />

or intermediary persons with specific marketing information<br />

and thus maximize the success of viral marketing campaigns.<br />

For example, the Procter & Gamble subsidiaries<br />

Vocal point and Tremor already use social network information<br />

to introduce new products, enjoying doubled sales<br />

in some test locations.<br />

Limitations and Directions for Further Research<br />

Although we designed our experiments carefully, some of<br />

their shortcomings might limit the validity of the results. In<br />

Study 1, order effects may exist because the sets of participants<br />

in the different experimental settings are not disjunctive.<br />

We designed Study 2 to avoid such an overlap, but<br />

the parallel timing in that experiment could lead to interrelationships<br />

across the different seeding strategies. However,<br />

as we noted previously, seeding strategies that lead<br />

to faster diffusion might have performed better, which also<br />

reflects reality well: In a world in which people become<br />

REFERENCES<br />

Anderson, Roy M. and Robert M. May (1991), Infectious Diseases<br />

of Humans: Dynamics and Control. New York: Oxford<br />

University Press.<br />

Aral, Sinan, Lev Muchnik, and Arun Sundararajan (2009),<br />

“Distinguishing Influence-Based Contagion from Homophily-<br />

<strong>Dr</strong>iven Diffusion in Dynamic Networks,” Proceedings of the<br />

National Academy of Sciences, 106 (51), 21544–49.<br />

, , and (2011), “Engineering Social<br />

Contagions: Optimal Network Seeding and Incentive Strate-<br />

increasingly exposed to multiple viral campaigns competing<br />

for their attention and participation, seeding strategies<br />

that lead to faster diffusion may be advantageous for the<br />

initiator of the particular viral marketing campaign.<br />

The student sample and the artificial information content<br />

of the experiments constitute additional limitations in<br />

our experimental studies, though these shortcomings do not<br />

systematically favor one strategy over another. It would be<br />

useful to conduct similar experiments with different samples<br />

and less artificial information, such as real-life viral<br />

marketing campaigns.<br />

In our real-life application, we could compare only highand<br />

low-degree and random seeding strategies. We did<br />

not differentiate referrals with regard to the profit of the<br />

referred customers (for an analysis of the value of customers<br />

acquired through referral programs, see Schmitt,<br />

Skiera, and Van den Bulte 2011), but our robustness checks<br />

show no significant regression-stage effects of additional<br />

covariates on converted reach or influence domain. Our<br />

results regarding the effects of degree centrality also remain<br />

unchanged across the robustness checks.<br />

Most current research focuses on individual choices and<br />

treats the choices of partners (within the social network)<br />

as exogenous; as we do, these studies assume that the network<br />

remains fixed for the duration of the study and unaffected<br />

by it. However, this strong assumption ignores the<br />

likely effects of dynamics inherent to real-life social networks<br />

(e.g., Steglich, Snijders, and Pearson 2010). Marketing<br />

response models that incorporate these effects would<br />

be a fruitful avenue for further research. Another extension<br />

might incorporate information on the dyadic level, such as<br />

tie strength, which could distinguish the success of referrals<br />

from a specific customer. In addition, from a marketing<br />

perspective, it would be useful to intensify research<br />

regarding the effects of incentives, as we find their impact<br />

to be quite significant in Study 1 (e.g., Schmitt, Skiera,<br />

and van den Bulte 2011; Aral, Munchnik, and Sundararajan<br />

2011).<br />

Our combination of two experimental studies and an<br />

ex post analysis of a real-world viral marketing campaign<br />

provides a strong argument that hubs and bridges are key<br />

to the diffusion of viral marketing campaigns. We cannot<br />

confirm recent findings that question the exposed role of<br />

hubs for the success of viral marketing campaigns, but our<br />

analysis of the different determinants in Study 3 yields<br />

additional explanations for why hubs make more attractive<br />

seeding points. These findings should serve as inputs to<br />

create more realistic computer simulations and analytical<br />

models.<br />

gies,” (accessed July 21, 2011), [available at http://ssrn.com/<br />

abstract=1770982].<br />

Arndt, Johan (1967), “Role of Product-Related Conversations<br />

in the Diffusion of a New Product,” Journal of Marketing<br />

Research, 4 (August), 291–95.<br />

Bampo, Mauro, Michael T. Ewing, Dineli R. Mather, David<br />

Stewart, and Mark Wallace (2008), “The Effects of the<br />

Social Structure of Digital Networks on Viral Marketing<br />

Performance,” Information Systems Research, 19 (3),<br />

273–90.<br />

Seeding Strategies for Viral Marketing / 69


Becker, Marshall H. (1970), “Sociometric Location and Innovativeness:<br />

Reformulation and Extension of the Diffusion Model,”<br />

American Sociological Review, 35 (2), 267–82.<br />

Berger, Jonah and Katy Milkman (2011), “Social Transmission,<br />

Emotion, and the Virality of Online Content,”<br />

(accessed July 21, 2011), [available at http://marketing.wharton<br />

.upenn.edu/documents/research/Virality.pdf].<br />

and Eric Schwartz (2011), “What Do People Talk About?<br />

<strong>Dr</strong>ivers of Immediate and Ongoing Word-of-Mouth,” Journal<br />

of Marketing Research, forthcoming.<br />

Biyalogorsky, Eyal, Eitan Gerstner, and Barak Libai (2001), “Customer<br />

Referral Management: Optimal Reward Programs,” Marketing<br />

Science, 20 (1), 82–95.<br />

Burt, Ronald S. (1987), “Social Contagion and Innovation: Cohesion<br />

Versus Structural Equivalence,”American Journal of Sociology,<br />

92 (6), 1287–1335.<br />

Coleman, James S., Elihu Katz, and Herbert Menzel (1966), Medical<br />

Innovation: A Diffusion Study. Indianapolis: Bobbs-Merrill.<br />

Costenbader, Elizabeth and Thomas W. Valente (2003), “The Stability<br />

of Centrality Measures When Networks Are Sampled,”<br />

Social Networks, 25 (4), 283–307.<br />

De Bruyn, Arnaud and Gary L. Lilien (2008), “A Multi-Stage<br />

Model of Word-of-Mouth Influence Through Viral Marketing,”<br />

International Journal of Research in Marketing, 25 (3), 151–63.<br />

Dodds, Peter S. and Duncan J. Watts (2004), “Universal Behavior<br />

in a Generalized Model of Contagion,” Physical Review Letters,<br />

92 (21), 1–5.<br />

Festinger, Leon (1957), A Theory of Cognitive Dissonance.<br />

Stanford, CA: Stanford University Press.<br />

Galeotti, Andrea and Sanjeev Goyal (2009), “Influencing the<br />

Influencers: A Theory of Strategic Diffusion,” RAND Journal<br />

of Economics, 40 (3), 509–532.<br />

Gladwell, Malcolm (2002), The Tipping Point: How Little Things<br />

Can Make a Big Difference. Boston: Back Bay Books.<br />

Godes, David and Dana Mayzlin (2009), “Firm-Created Word-of-<br />

Mouth Communication: Evidence from a Field Study,” Marketing<br />

Science, 28 (4), 721–39.<br />

Goldenberg, Jacob, Sangman Han, Donald R. Lehmann, and Jae<br />

Weon Hong (2009), “The Role of Hubs in the Adoption Process,”<br />

Journal of Marketing, 73 (2), 1–13.<br />

Granovetter, Mark S. (1973), “The Strength of Weak Ties,” American<br />

Journal of Sociology, 78 (6), 1360–80.<br />

(1985), “Economic Action and Social Structure: The<br />

Problem of Embeddedness,” American Journal of Sociology, 91<br />

(3), 481–510.<br />

Hanaki, Nobuyuki, Alexander Peterhansl, Peter S. Dodds, and<br />

Duncan J. Watts (2007), “Cooperation in Evolving Social Networks,”<br />

Management Science, 53 (7), 1036–50.<br />

Hann, Il-Horn, Kai-Lung Hui, Sang-Yong T. Lee, and Ivan P.L.<br />

Png (2008), “Consumer Privacy and Marketing Avoidance:<br />

A Static Model,” Management Science, 54 (6), 1094–1103.<br />

Hill, Shawndra, Foster Provost, and Chris Volinsky (2006),<br />

“Network-Based Marketing: Identifying Likely Adopters via<br />

Consumer Networks,” Statistical Science, 21 (2), 256–76.<br />

Hinz, Oliver and Martin Spann (2008), “The Impact of Information<br />

Diffusion on Bidding Behavior in Secret Reserve Price<br />

Auctions,” Information Systems Research, 19 (3), 351–68.<br />

Iyengar, Raghuram, Christophe Van den Bulte, and Thomas W.<br />

Valente (2011), “Opinion Leadership and Social Contagion in<br />

New Product Diffusion,” Marketing Science, 30 (2), 195–212.<br />

Iyengar, Sheena S. and Mark Lepper (2000), “When Choice<br />

Is Demotivating: Can One Desire Too Much of a Good<br />

Thing?” Journal of Personality and Social Psychology, 76 (6),<br />

995–1006.<br />

70 / Journal of Marketing, November 2011<br />

Jain, Dipak, Vijay Mahajan, and Eitan Muller (1995), “An<br />

Approach for Determining Optimal Product Sampling for the<br />

Diffusion of a New Product,” Journal of Product Innovation<br />

Management, 12 (2), 124–35.<br />

Kalish, Shlomo, Vijay Mahajan, and Eitan Muller (1995), “Waterfall<br />

and Sprinkler New-Product Strategies in Competitive<br />

Global Markets,” International Journal of Research in Marketing,<br />

12 (2), 105–119.<br />

Kemper, John T. (1980), “On the Identification of Superspreaders<br />

for Infectious Disease,” Mathematical Biosciences, 48 (1/2),<br />

111–27.<br />

Kiss, Christine and Martin Bichler (2008), “Identification of<br />

Influencers—Measuring Influence in Customer Networks,”<br />

Decision Support Systems, 46 (1), 233–53.<br />

Lambert, Diane (1992), “Zero-Inflated Poisson Regression, with<br />

an Application to Defects in Manufacturing,” Technometrics,<br />

34 (1), 1–14.<br />

Leskovec, Jure, Lada A. Adamic, and Bernardo A. Huberman<br />

(2007), “The Dynamics of Viral Marketing,” ACM Transactions<br />

on the Web, 1 (1), 228–37.<br />

Libai, Barak, Eitan Muller, and Renana Peres (2005), “The Role<br />

of Seeding in Multi-Market Entry,” International Journal of<br />

Research in Marketing, 22 (4), 375–93.<br />

Lin, Nan (1976), Foundations of Social Research. New York:<br />

McGraw-Hill.<br />

Mullahy, John (1986), “Specification and Testing of Some Modified<br />

Count Data Models,” Journal of Econometrics, 33 (3),<br />

341–65.<br />

Porter, Constance E. and Naveen Donthu (2008), “Cultivating<br />

Trust and Harvesting Value in Virtual Communities,” Management<br />

Science, 54 (1), 113–28.<br />

Porter, Lance and Guy J. Golan (2006), “From Subservient Chickens<br />

to Brawny Men: A Comparison of Viral Advertising to<br />

Television Advertising,” Journal of Interactive Advertising, 6<br />

(2), 30–38.<br />

Rayport, Jeffrey (1996), “The Virus of Marketing,” Fast Company<br />

(accessed March 24, 2011), [available at http://www<br />

.fastcompany.com/magazine/06/virus.html].<br />

Schmitt, Philipp, Bernd Skiera, and Christophe Van den Bulte<br />

(2011), “Referral Programs and Customer Value,” Journal of<br />

Marketing, 75 (January), 46–59.<br />

Scott, John (2000), Social Network Analysis: A Handbook, 2d ed.<br />

London: Sage Publications.<br />

Simmel, Georg (1950), The Sociology of Georg Simmel, compiled<br />

and translated by Kurt Wolff. Glencoe, IL: The Free Press.<br />

Steglich, Christian, Tom A.B. Snijders, and Michael Pearson<br />

(2010), “Dynamic Networks and Behavior: Separating<br />

Selection from Influence,” Sociological Methodology, 40 (1),<br />

329–93.<br />

Sundararajan, Arun (2006), “Network Seeding,” Extended<br />

Abstract for WISE 2006, New York University Stern School of<br />

Business.<br />

Thompson, Clive (2008), “Is the Tipping Point Toast?” Fast<br />

Company, 122, 74–105, (accessed March 24, 2011), [available<br />

at http://www.fastcompany.com/magazine/122/is-the-tippingpoint-toast.html].<br />

Toubia, Olivier, Andrew T. Stephen, and Aliza Freud (2010),<br />

“Identifying Active Members in Viral Marketing Campaigns,”<br />

(accessed July 21, 2011), [available at http://sites.google.com/a/<br />

andrewstephen.net/andrew/researc/papers/Toubia_Stephen_Freud_<br />

viralmarketing_nov2010.pdf].<br />

Trusov, Michael, Anand Bodapati, and Randolph E. Bucklin<br />

(2010), “Determining Influential Users in Internet Social Networks,”<br />

Journal of Marketing Research, 47 (August), 643–58.


Van den Bulte, Christophe (2010), “Opportunities and Challenges<br />

in Studying Consumer Networks,” in The Connected Customer,<br />

S. Wuyts, M.G. Dekimpe, E. Gijsbrechts, and R. Pieters, eds.<br />

London: Routledge, 7–35.<br />

and Yogesh V. Joshi (2007), “New Product Diffusion<br />

with Influentials and Imitators,” Marketing Science, 26 (3),<br />

400–421.<br />

and Gary L. Lilien (2001), “Medical Innovation Revisited:<br />

Social Contagion Versus Marketing Effort,” American<br />

Journal of Sociology, 106 (5), 1409–1435.<br />

and Stefan Wuyts (2007), Social Networks and Marketing.<br />

Cambridge, MA: MSI Relevant Knowledge Series.<br />

Van der Lans, <strong>Ralf</strong>, Gerrit van Bruggen, Jehoshua Eliashberg,<br />

and Berend Wierenga (2010), “A Viral Branching Model for<br />

Predicting the Spread of Electronic Word-of-Mouth,” Marketing<br />

Science, 29 (2), 348–65.<br />

Watts, Duncan J. (2004), “The ‘New’ Science of Networks,”<br />

Annual Review of Sociology, 30, 243–70.<br />

and Peter S. Dodds (2007), “Influentials, Networks, and<br />

Public Opinion Formation,” Journal of Consumer Research, 34<br />

(4), 441–58.<br />

and Jonah Peretti (2007), “Viral Marketing for the Real<br />

World,” Harvard Business Review, 85 (5), 22–23.<br />

Seeding Strategies for Viral Marketing / 71


© 2011, American Marketing Association<br />

ISSN: 0022-2429 (print), 1547-7185 (electronic)<br />

Bernd Skiera, Manuel Bermes, & Lutz Horn<br />

Customer Equity Sustainability<br />

Ratio: A New Metric for Assessing a<br />

Firm’s Future Orientation<br />

Securitization is a remarkable financial instrument; it enables securitizers to increase their short-term profits at the<br />

expense of the long-term value of their customer base. This ability might be tempting for firms, especially because<br />

it does not need to be disclosed transparently to stakeholders. The authors show how their newly developed<br />

customer equity sustainability ratio (CESR) complements customer equity reporting and creates more transparency<br />

about the consequences of securitization for future earnings and the riskiness of the underlying business model.<br />

Their model compares the future value of an existing customer base with current earnings. In an empirical study<br />

of 38 banks in ten countries, the authors demonstrate the limited transparency of long-term value creation in<br />

financial statements. Next, they outline the adequacy of CESR for creating more transparency in empirical cases<br />

of Countrywide Financial Corporation and nine firms from nonbanking industries. They recommend that marketing<br />

should play a prominent role in providing stakeholders with substantial information about the long-term value of the<br />

customer base.<br />

Keywords: customer equity, financial reporting, sustainability, securitization, financial crisis, customer equity<br />

sustainability ratio<br />

The financial crisis currently disrupting the economic<br />

system and banking worldwide is often attributed to<br />

securitization (Ryan 2008), an instrument that pools<br />

assets with an inherent stream of future earnings to support<br />

debt instruments, called asset-backed securities (ABSs), for<br />

sale to investors (Greenbaum and Thakor 1987). 1 Banks<br />

commonly use securitization to manage their portfolio risk<br />

and funding position by transferring loans and the credit<br />

risk of their loan portfolios to other investors (Santomero<br />

and Babbel 1997). In return, the banks do not hold the loans<br />

until maturity in their own books but instead receive earnings<br />

from them directly at their present value (PV).<br />

Unfortunately, as the global financial crisis makes clear,<br />

the inherent risk of ABSs is difficult to calculate because of<br />

their frequent repackaging; most market participants<br />

grossly underestimated it (Coval, Jurek, and Stafford 2009).<br />

1We use ABS to refer to the whole group of structured products,<br />

which can be classified into asset-backed commercial papers,<br />

mortgage-backed securities, and collateralized debt obligations.<br />

Bernd Skiera is Chair of Electronic Commerce, Department of Marketing,<br />

Faculty of Business and Economics (e-mail: skiera@skiera.de), Manuel<br />

Bermes is a doctoral student, E-<strong>Finance</strong> <strong>Lab</strong>, House of <strong>Finance</strong> (e-mail:<br />

bermes@ wiwi.uni-frankfurt.de), and Lutz Horn is a doctoral student,<br />

Retail Banking Competence Center, House of <strong>Finance</strong> (e-mail: lutzhorn@<br />

gmx.de), Goethe University Frankfurt. The authors thank Andreas Hackethal,<br />

Clemens Jochum, <strong>Wolfgang</strong> <strong>König</strong>, Jan Pieter Krahnen, Christian<br />

Leuz, Steffen Meyer, Anita Mosch, Philipp Schmitt, Christian Schulze,<br />

Thorsten Wiesel, seminar participants at the University of Technology<br />

Sydney and Maastricht University, and three anonymous JM reviewers for<br />

their valuable comments on earlier drafts of the article.<br />

118<br />

As a consequence, prices for securities dropped dramatically,<br />

banks were forced to realize heavy write-downs on<br />

their asset bases, and they suffered huge losses and in some<br />

cases bankruptcies (for more detailed descriptions, see<br />

Demyanyk and Van Hemert 2009; Franke and Krahnen<br />

2008).<br />

These risks are well known, if not fully solved; however,<br />

additional problems result from moves toward short-term<br />

profit realizations that are inherent to the securitization of<br />

loans and come at the expense of long-term value creation.<br />

The transformation of periodic loan payments into one<br />

down payment enables a bank to realize the PV of earnings<br />

immediately rather than doing so over the lifetime of the<br />

loan.<br />

Securitization is not limited to the banking industry.<br />

Airlines and sport clubs can securitize earnings from ticket<br />

sales that will occur over the next few years to realize the<br />

earnings in the present rather than spreading them over<br />

time. Although it remains particularly prevalent in banking,<br />

securitization enjoys great popularity in other industries<br />

such as telecommunications, utilities, and the music business;<br />

it is even used by national governments and other public<br />

institutions (Brinkworth 2004; Downey 1999; Ketkar<br />

and Ratha 2008).<br />

Supported by accounting rules, managers do not need to<br />

make the consequences of securitization for long-term<br />

value creation transparent. They have incentives to adjust<br />

their firm’s earnings stream through securitization and<br />

reach their own goals, such as greater personal wealth<br />

(Dechow and Shakespeare 2009). However, the problems<br />

that arise from such adjustments in customer management<br />

Journal of Marketing<br />

Vol. 75 (May 2011), 118 –131


strategies and earnings streams continue to be largely<br />

ignored in current discussions of the financial crisis (e.g.,<br />

Coval, Jurek, and Stafford 2009; Franke and Krahnen<br />

2008). Not surprisingly, metrics that allow for easily detecting<br />

shifts in earning streams are scarce.<br />

This article outlines the problems associated with this<br />

shift in earnings and proposes customer equity reporting<br />

(CER) along with a new ratio, the customer equity sustainability<br />

ratio (CESR), as means to increase the level of transparency<br />

in financial statements. We emphasize the importance<br />

of reporting forward-looking marketing metrics in<br />

financial statements, thereby extending the role of marketing:<br />

It should supply firm stakeholders with substantial<br />

information about the long-term value of the current customer<br />

base (see also Joshi and Hanssens 2010; Tuli,<br />

Bharadwaj, and Kohli 2010). In line with Wiesel, Skiera, and<br />

Villanueva’s (2008) recent proposal, which postulates that<br />

CER can complement financial statements, we argue that<br />

greater transparency, achieved by reporting more forwardlooking<br />

marketing metrics, might have reduced the devastating<br />

consequences of the current financial crisis for banks<br />

and might lead to a suitable use of securitization in industries<br />

outside banking.<br />

We begin by describing securitization, its advantages,<br />

and its disadvantages. Next, we show how CER and CESR<br />

can capture some of the effects of securitization previously<br />

neglected by first focusing on banking business and its<br />

extensive use of securitization. Then, we empirically analyze<br />

the transparency of long-term value creation in the<br />

financial statements of 38 banks in ten countries and apply<br />

our reporting technique, including CESR, to the former<br />

U.S. market leader in mortgage lending and origination,<br />

Countrywide Financial Corporation. We also explore industries<br />

outside banking and show, for nine securitizations of<br />

firms and institutions in diverse industries, how CESR can<br />

detect differences in the extent of shifts in earnings across<br />

time. We conclude with a discussion of the results.<br />

Securitization<br />

Basic Idea of Securitization<br />

Securitization is best known as the pooling and repackaging<br />

of a group of assets (e.g., loans) and the subsequent sale of<br />

tranches, which are the new divisions of the group of assets<br />

classified by asset quality, of this pool to new investors<br />

(Ryan 2008; Santomero and Babbel 1997). For example,<br />

banks commonly sell a variety of financial products, from<br />

mortgages to student loans to credit cards to leasing claims,<br />

which provide the underlying loans for their securitization<br />

(Ketkar and Ratha 2008; Santomero and Babbel 1997).<br />

They transfer the loan assets for regulatory and accounting<br />

purposes into a separate business unit, called a special purpose<br />

vehicle (SPV), and group them into tranches (Greenbaum<br />

and Thakor 1987). Each tranche receives a rating<br />

from a rating agency, from senior (AAA to A rating) to<br />

equity (no rating/residuum). Then, these tranches can be<br />

priced and sold to new investors (Coval, Jurek, and Stafford<br />

2009; Luo, Tang, and Wang 2009), who receive all earnings<br />

from the tranche they own, which means they also confront<br />

any risks arising from the underlying loans (see Franke and<br />

Krahnen 2008).<br />

Not just banks and other financial institutions but virtually<br />

all firms and even state and national governments can<br />

and do use securitization. The main prerequisite to sell<br />

claims to investors is an underlying asset that currently is<br />

generating earnings or will generate them in the future (for<br />

further details, see Kendall 1996; Kothari 2006). A wide<br />

variety of assets from different industries can be used for<br />

securitization, such as future revenues from cellular phone<br />

contracts or electricity consumption, ticket sales from<br />

future soccer games, airline ticket sales, tax revenue receivables,<br />

and royalties of intellectual properties (Brinkworth<br />

2004; Downey 1999; Ketkar and Ratha 2008). Prominent<br />

recent securitization transactions include musicians such as<br />

David Bowie, who issued the first music royalties future<br />

receivables securitization (Burke Sylva 1999); soccer teams<br />

such as Leeds United (United Kingdom), Tottenham Hotspurs<br />

(United Kingdom), and Schalke 04 (Germany), which<br />

have used securitization to fund their entire businesses<br />

(Brinkworth 2004); and U.S. football teams such as the<br />

Denver Broncos, which financed a new stadium by using<br />

securitization “backed by about 4,000 stadium-related contracts,<br />

such as luxury box seats, club seats, a portion of concession<br />

fees and other cash flows” (Gregory 2002, p. 6).<br />

Some banks helped Greece improve its short-term financial<br />

situation and hide its debt level by providing cash in return<br />

for Greece’s government payments in the future, meaning<br />

that Greece has traded away its revenue from the rights of<br />

airport fees and lotteries (Story, Thomas, and Schwartz<br />

2010).<br />

Numerical Example to Outline the Effects of<br />

Securitization<br />

Despite the widespread popularity of securitization, we<br />

concentrate on banks because loans provide a clear illustration<br />

of its effects. We depict a loan example to show the<br />

effects of securitization on the earnings stream (for comparable<br />

arguments, see Fabozzi, Davis, and Choudry 2006). In<br />

this example, we use a bank that issues one five-year loan<br />

volume of $100,000 to customers at the beginning of each<br />

year, with an annual interest rate of 5% paid at the end of<br />

each year. The customers repay the loan linearly over its<br />

lifetime, so the effective loan volume reduces to $80,000 in<br />

Year 2, $60,000 in Year 3, and so on. During the lifetime of<br />

the loan, the bank receives annual interest income of $5,000<br />

in Year 1, $4,000 in Year 2, $3,000 in Year 3, and so on. We<br />

assume an interest rate of 3.5% for the financial debt to refinance<br />

the loan issuance, resulting in interest expenses of<br />

$3,500 in Year 1, $2,800 in Year 2, $2,100 in Year 3, and so<br />

on. We also deduct loan loss provisions, equal to .5% of the<br />

loan volume, to account for potential default of these customers.<br />

So the bank realizes a margin of 1% (5% – 3.5% –<br />

.5%) of the loan volume. Although specific terms of loans,<br />

such as prepayments, deductions, servicing, and other costs,<br />

might increase the complexity of this illustration in practice,<br />

we avoid additional complexity for ease of exposition.<br />

We also assume that annual redemptions repay the corresponding<br />

financial debt.<br />

Customer Equity Sustainability Ratio (CESR) / 119


Table 1 includes the results for a non-securitizing bank,<br />

and Table 2 depicts the outcome for a securitizing bank. To<br />

show the time-shifting effects of securitization on the earnings<br />

of the bank in the second part of our example, we<br />

assume that from Year 8 onward, the bank does not issue<br />

any new loans and the existing loans expire. Until Year 7,<br />

expiring loans are permanently renewed by new issuances,<br />

and for ease of exposition we assume that the bank has<br />

reached a steady state, so that its total loan volume remains<br />

stable. The earnings of $3,000 in Year 1 comprise $1,000<br />

from loans issued in Year 1, $800 from loans of the prior<br />

year (Year 0), $600 from the next to last year (Year –1),<br />

$400 from the remaining loans of Year –2, and $200 from<br />

those of Year –3. In Year 2, the bank issues new loans generating<br />

earnings of $1,000, earnings of loans from Year –2<br />

to Year 1 decline by $200 each, and loans from Year –3<br />

expire.<br />

The bank can either distribute the obtained earnings as<br />

dividends to its shareholders (distribution case) or keep and<br />

reinvest them (reinvestment case). In the distribution case,<br />

the initial equity of $30,000 remains stable over time.<br />

Because earnings also are constant, the bank realizes a<br />

return on equity (ROE) of 10% each year. In the reinvestment<br />

case, we assume an annual return rate of the reinvestment<br />

of 10%, equivalent to the discount rate. The bank reinvests<br />

its earnings each year for a one-year period,<br />

repeated until the end of the loan contract, and does not distribute<br />

any earnings, so the earnings from previous years are<br />

permanently reinvested. For example, in Year 3, the bank<br />

can reinvest the sum ($6,300) of the earnings of the previous<br />

year ($3,000); the earnings of this year ($3,000), which<br />

are labeled as new reinvestment; as well as the earnings<br />

from the reinvestment in Year 2 (10% ¥ $3,000 = $300). Its<br />

equity increases from $33,000 ($30,000 + $3,000) to<br />

$36,300 because of the new reinvestment ($3,000) and the<br />

earnings from reinvestment ($300). Thus, the bank’s starting<br />

equity of $30,000 grows every year from the retained<br />

earnings from loans and reinvestment. The ROE still<br />

amounts to 10% because the relative increase in earnings<br />

equals the rise in equity.<br />

We also consider a securitizing bank (Table 2) that<br />

decides, in Year 3, to transfer the whole loan volume and<br />

related earnings to new investors. 2 We assume the bank<br />

does so at fair market value, or the PV of earnings at the<br />

end of Year 3. Thus, the bank receives the earnings from the<br />

loan securitization of $6,397. This amount reflects the PV<br />

of new loans issued in Year 3, or $2,660, plus the PV of the<br />

remaining loans of Years –1 to 2, equal to $3,737 ([$800 +<br />

$600 + $400 + $200] + [$600 + $400 + $200]/1.1 + [$400 +<br />

$200]/1.1 2 + $200/1.1 3), that were securitized in Year 3 as<br />

well. However, in considering the PV of future years, the<br />

securitizing bank realizes only $2,660, equal to the PV of<br />

the new loans acquired and securitized each year ($1,000 +<br />

$800/1.1 + $600/1.1 2 + $400/1.1 3 + $200/1.1 4).<br />

2We assume a direct origination and distribution of loans to new<br />

investors. However, an indirect sale through a bank-owned SPV,<br />

which is a separate legal entity, is common as a means to obtain<br />

various accounting reliefs, though it has no influence on the results<br />

in our example.<br />

120 / Journal of Marketing, May 2011<br />

We again distinguish between distributing and reinvesting<br />

the bank’s earnings. In the distribution case, equity<br />

remains constant at $30,000. The securitizing bank reaches<br />

an ROE of 10% in Years 1 and 2, which increases to 21.3%<br />

in Year 3 because of the shift to short-term profit realization<br />

through securitization. However, the increase in ROE in<br />

Year 3 comes at the expense of a lower ROE from Year 4<br />

onward; it falls back to a steady-state level of 8.9%.<br />

In Year 3, with earnings of $6,397, the securitizing bank<br />

surpasses the earnings of the non-securitizing bank by more<br />

than 100%. After Year 3, the securitizing bank earns only<br />

$2,660, whereas the non-securitizing bank continues to earn<br />

$3,000 per year, with more future earnings already contracted.<br />

Comparing ROE leads to a similar result: The nonsecuritizing<br />

bank shows a stable ROE of 10%, so the securitizing<br />

bank outperforms it in Year 3 with a ROE of 21.3%<br />

but then realizes only 8.9% from Year 4 onward.<br />

In the reinvestment case, equity increases by the earnings<br />

realized from loans, securitization, and reinvestment.<br />

When securitizing in Year 3, ROE grows to 19.4% because<br />

of the shift from long-term value creation to short-term<br />

profit realization. However, one year later, ROE falls back<br />

to a lower level of 9.2% and rises only slowly to 9.4% in<br />

Year 7. Compared with the distribution case, the differences<br />

in ROE result from the reinvestment of earnings and the<br />

compound interest effects of this reinvestment.<br />

In this example, both the seller (i.e., securitizer) and the<br />

buyer of the loans have equal assessments of the underlying<br />

(risky) earnings and use the same discount rate, and the<br />

seller does not realize a higher return rate than the discount<br />

rate through reinvestment. Thus, securitization neither creates<br />

nor destroys value but again simply shifts the realization<br />

of value over time (Kothari 2006). Differences in value<br />

accrue only when the seller and buyer differ in (1) their<br />

assessments of the underlying (risky) earnings or (2) their<br />

discount rates, because (1) and (2) yield to different PVs, or<br />

(3) the seller can use the earnings brought forward for reinvestment<br />

to realize a return rate greater than the discount<br />

rate. Still, securitizers increase their short-term profits at the<br />

expense of long-term value creation.<br />

Treatment of Securitization by Reporting<br />

Standards<br />

Reporting standards provide few opportunities to detect the<br />

earnings shift from long-term value creation toward shortterm<br />

profit generation because firms are not required to<br />

report quantitative outlooks of their future earnings. The<br />

U.S. Financial Accounting Standards Board (FASB) introduced<br />

FAS 125 and FAS 140, which obligate firms to publish<br />

more detailed information about their securitization<br />

transactions and valuation principles (FASB 1996, 2000;<br />

Ryan 2008); comparable International Financial Reporting<br />

Standards call for less information about securitization in<br />

IAS 32 and 39 (McCreevy 2008). Therefore, current regulations<br />

require banks and firms, respectively, to publish little<br />

information about their securitization activities; they might<br />

disclose securitization volume in the amendments of their<br />

financial statements, but it is not mandatory. The earnings<br />

derived from securitization may be registered as noninterest<br />

or interest income, which do not need to be separated from


TAblE 1<br />

Numerical Example of loans of a Non-Securitizing bank<br />

Non-Securitizing bank Year 1 Year 2 Year 3 Year 4 Year 5 Year 6 Year 7 Year 8 Year 9 Year 10 Year 11<br />

Noninterest income $0 $0 $0 $0 $0 $0 $0 $0 $0 $0 $0<br />

Net interest income from<br />

Loans issued in Year –3 $200<br />

Loans issued in Year –2 $400 $200<br />

Loans issued in Year –1 $600 $400 $200<br />

Loans issued in Year 0 $800 $600 $400 $200<br />

Loans issued in Year 1 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 2 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 3 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 4 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 5 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 6 $1,000 $800 $600 $400 $200<br />

Loans issued in Year 7 $1,000 $800 $600 $400 $200<br />

Net interest income $3,000 $3,000 $3,000 $3,000 $3,000 $3,000 $3,000 $2,000 $1,200 $600 $200<br />

Earnings $3,000 $3,000 $3,000 $3,000 $3,000 $3,000 $3,000 $2,000 $1,200 $600 $200<br />

Distribution Case<br />

Customer equity (as of 12/31) $6,397 $6,397 $6,397 $6,397 $6,397 $6,397 $6,397<br />

Customer equity sustainability<br />

ratio (as of 12/31) .531 .531 .531 .531 .531 .531 .531<br />

Loan volume to customers<br />

(at beginning of year) $300,000 $300,000 $300,000 $300,000 $300,000 $300,000 $300,000<br />

Equity $30,000 $30,000 $30,000 $30,000 $30,000 $30,000 $30,000<br />

Return on equity 10.0% 10.0% 10.0% 10.0% 10.0% 10.0% 10.0%<br />

Reinvestment Case<br />

Earnings from reinvestment $0 $300 $630 $993 $1,392 $1,832 $2,315 $2,846 $3,331 $3,784 $4,222<br />

New reinvestment –$3,000 –$3,300 –$3,630 –$3,993 –$4,392 –$4,832 –$5,315 –$4,846 –$4,531 –$4,384 $42,222<br />

Total result from reinvestment –$3,000 –$3,000 –$3,000 –$3,000 –$3,000 –$3,000 –$3,000 –$2,000 –$1,200 –$600 $46,445<br />

Customer equity (as of 12/31) $6,670 $7,270 $7,930 $8,656 $9,455 $10,333 $11,299<br />

Customer equity sustainability<br />

ratio (as of 12/31) .550 .546 .542 .539 .535 .532 .530<br />

Loan volume to customers<br />

(at beginning of year) $300,000 $300,000 $300,000 $300,000 $300,000 $300,000 $300,000<br />

Equity $30,000 $33,000 $36,300 $39,930 $43,923 $48,315 $53,147<br />

Return on equity 10.0% 10.0% 10.0% 10.0% 10.0% 10.0% 10.0%<br />

Customer Equity Sustainability Ratio (CESR) / 121


TAblE 2<br />

Numerical Example of loans of a Securitizing bank<br />

Securitizing bank Year 1 Year 2 Year 3 Year 4 Year 5 Year 6 Year 7 Year 8 Year 9 Year 10 Year 11<br />

Noninterest income $0 $0 $6,397 $2,660 $2,660 $2,660 $2,660 $0 $0 $0 $0<br />

Net interest income from<br />

Loans issued in Year –3 $200<br />

Loans issued in Year –2 $400 $200<br />

Loans issued in Year –1 $600 $400<br />

Loans issued in Year 0 $800 $600<br />

Loans issued in Year 1 $1,000 $800<br />

Loans issued in Year 2 $1,000<br />

Loans issued in Year 3<br />

Loans issued in Year 4<br />

Loans issued in Year 5<br />

Loans issued in Year 6<br />

Loans issued in Year 7<br />

Net interest income $3,000 $3,000 $0 $0 $0 $0 $0 $0 $0 $0 $0<br />

Earnings $3,000 $3,000 $6,397 $2,660 $2,660 $2,660 $2,660 $0 $0 $0 $0<br />

122 / Journal of Marketing, May 2011<br />

Distribution Case<br />

Customer equity (as of 12/31) $6,397 $6,397 $6,397 $2,660 $2,660 $2,660 $2,660<br />

Customer equity sustainability<br />

ratio (as of 12/31) .531 .531 .000 .000 .000 .000 .000<br />

Loan volume to customers<br />

(at beginning of year) $300,000 $300,000 $0 $0 $0 $0 $0<br />

Equity $30,000 $30,000 $30,000 $30,000 $30,000 $30,000 $30,000<br />

Return on equity 10.0% 10.0% 21.3% 8.9% 8.9% 8.9% 8.9%<br />

Reinvestment Case<br />

Earnings from reinvestment $0 $300 $630 $1,333 $1,732 $2,171 $2,654 $3,186 $3,504 $3,855 $4,240<br />

New reinvestment –$3,000 –$3,300 –$7,027 –$3,993 –$4,392 –$4,832 –$5,315 –$3,186 –$3,504 –$3,855 $42,404<br />

Total result from reinvestment –$3,000 –$3,000 –$6,397 –$2,660 –$2,660 –$2,660 –$2,660 $0 $0 $0 $46,645<br />

Customer equity (as of 12/31) $6,670 $7,270 $8,239 $5,568 $6,366 $7,245 $8,211<br />

Customer equity sustainability<br />

ratio (as of 12/31) .550 .546 .147 .283 .310 .333 .353<br />

Loan volume to customers<br />

(at beginning of year) $300,000 $300,000 $0 $0 $0 $0 $0<br />

Equity $30,000 $33,000 $36,300 $43,327 $47,320 $51,713 $56,544<br />

Return on equity 10.0% 10.0% 19.4% 9.2% 9.3% 9.3% 9.4%


other noninterest or interest income positions. Therefore, the<br />

shift from creating long-term value to realizing short-term<br />

profits can be hidden well (Dechow and Shakespeare 2009).<br />

Advantages and Disadvantages of Securitization<br />

Securitization offers several major advantages to firms (for an<br />

extended discussion, see Franke and Krahnen 2008). First,<br />

securitization helps banks manage the various risks of underlying<br />

loan portfolios. They can construct well-diversified<br />

investments and lower or even avoid risks by transferring<br />

them to other investors. In risk transfers, banks might also<br />

be able to realize profits if the sellers and buyers of loans<br />

maintain different evaluations of the fair value. Second,<br />

firms can attain stronger capital, funding, and liquidity positions<br />

through securitization because it enables them to turn<br />

illiquid assets, such as mortgages, into more liquid assets.<br />

Third, securitization supports banks in their efforts to fulfill<br />

regulatory requirements such as Basel II, because they no<br />

longer need to hold equity for securitized loans (Calem and<br />

LaCour-Little 2004). Fourth, the lower equity requirements<br />

for securitized loans increase ROE.<br />

However, securitization also produces two major dis -<br />

advantages. First, the frequent repackaging and splitting of<br />

ABSs into tranches make it difficult to determine the appropriate<br />

value of those securities. This topic appears in extensive<br />

discussions elsewhere, so we do not detail it here (see<br />

Coval, Jurek, and Stafford 2009; Luo, Tang, and Wang<br />

2009). Second, the short- and long-term effects of securitization<br />

are not equally transparent. The result of the shortterm<br />

effect, usually an increase in earnings, is clearly<br />

described in financial reports, but the corresponding<br />

decrease in long-term value is less obvious and, as we show<br />

in our empirical study, is not well reported by the vast<br />

majority of firms. The low transparency causes various<br />

problems, including those related to management compensation<br />

when variable payments are linked to short-term<br />

profits (Dechow, Myers, and Shakespeare 2010; Dechow<br />

and Shakespeare 2009).<br />

Customer Equity Reporting and<br />

Securitization<br />

Customer Equity Reporting<br />

Customer equity reporting creates the required transparency<br />

by reporting the value of the customer base (i.e., customer<br />

equity) and its development over time. Blattberg and<br />

Deighton (1996) define “customer equity” (CE) as the customer<br />

lifetime values (CLV) of the firm’s current customers j:<br />

( 1)<br />

CE = CLVj. j = 1<br />

If the total lifespan of a customer j is T j, CLV is simply the<br />

PV of customer j’s earnings (Earn j,t) over time t (discounted<br />

at rate i):<br />

J<br />

∑<br />

Earn j, t<br />

( 2)<br />

CLVj<br />

= .<br />

t ( 1 + i)<br />

T j<br />

∑<br />

t = 0<br />

Blattberg and Deighton (1996), Gupta, Lehmann, and<br />

Stuart (2004), Rust, Lemon, and Zeithaml (2004), and Rust,<br />

Zeithaml, and Lemon (2000) lay a foundation for a deeper<br />

understanding and wider acceptance of the CE approach in<br />

science and practice (see also Srinivasan and Hanssens<br />

2009). As Wiesel, Skiera, and Villanueva (2008) indicate,<br />

CER extends this field of application by connecting CE to<br />

financial reporting. We outline an opportunity to define CE<br />

more narrowly than Wiesel, Skiera, and Villanueva do for<br />

firms in industries with high shares of contracted business,<br />

such as banking, telecommunications, and electricity. In<br />

this contractual setting, in general, customer relationships<br />

span several years, and future earnings can be projected<br />

reliably because they are contractually assured. We consider<br />

only contracted future earnings for our calculation of CLV<br />

and CE (unless noted otherwise).<br />

Customer Equity Sustainability Ratio<br />

We propose the CESR as a new ratio to quantify the intensity<br />

of long-term value creation and establish a connection<br />

between a firm’s financial statements and forward-looking<br />

CER. The CESR contrasts the future value of an existing<br />

customer base with current earnings, such that for an individual<br />

customer j, the future value is the PV of all earnings<br />

after the current year t = 0. Defining CESRj as the ratio of<br />

the PV of all future earnings to the corresponding CLV and<br />

rearranging leads to the following:<br />

( 3)<br />

CESR<br />

j<br />

=<br />

Tj<br />

∑<br />

t = 1<br />

Tj<br />

∑<br />

t = 0<br />

Earn j, t<br />

t ( 1 + i)<br />

CLV − Earn<br />

=<br />

Earn<br />

CLVj<br />

j, t<br />

t<br />

( 1 + i)<br />

Earn j<br />

= 1 − .<br />

CLV<br />

j j,<br />

0 , 0<br />

Therefore, the CESR for all current customers is<br />

( 4)<br />

CESR =<br />

J<br />

Tj<br />

∑∑<br />

j = 1 t = 1<br />

J<br />

Tj<br />

∑∑<br />

j = 1 t = 0<br />

Earn<br />

j, t<br />

t<br />

( 1 + i)<br />

=<br />

Earn<br />

j, t<br />

t<br />

( 1 + i)<br />

= 1 −<br />

J<br />

∑<br />

j = 1<br />

J<br />

∑<br />

j = 1<br />

j<br />

j = 1 j = 1<br />

J<br />

j = 1<br />

If we define current earnings as Earn 0 = S J j = 1Earn j,0, we<br />

can rewrite Equation 4 as follows:<br />

In the case of nonnegative earnings and positive CE, the<br />

CESR falls between 0 and 1. A higher CESR indicates that<br />

the future value of the current customer base is high. A<br />

CESR of 0 implies that all earnings occur in the current<br />

year, as in our securitization example, whereas a CESR of 1<br />

J<br />

∑ ∑<br />

Earn<br />

CLV<br />

CLV − Earn<br />

j,<br />

0<br />

∑<br />

Earn0<br />

( 5) CESR = 1 −<br />

.<br />

CE<br />

j,<br />

0<br />

Customer Equity Sustainability Ratio (CESR) / 123<br />

j<br />

.<br />

J<br />

CLV<br />

j<br />

j


presumes that all earnings will be realized in the future,<br />

with no current earnings.<br />

Numerical CER and CESR Example<br />

To explain the relevance of CER and CESR in more detail,<br />

we adapt the concept to our numerical example in Tables 1<br />

and 2. The non-securitizing bank keeps issuing loans at an<br />

annual volume of $100,000 and realizes earnings of $3,000<br />

from the loans each year. In the distribution case, annual CE<br />

is $6,397, or the PV of all current ($3,000) and future,<br />

already contracted earnings ($3,397 = $2,000/1.1 + $1,200/<br />

1.12 + $600/1.13 + $200/1.14). The CESR is .531 (1 –<br />

$3,000/ $6,397). The bank receives 47% ($3,000/$6,397) of<br />

its contracted earnings in the current year and 53% ($3,397/<br />

$6,397) in the future. In the securitization case, the bank<br />

realizes all future earnings (from Year 3 onward) immediately,<br />

so CE from Year 4 onward decreases to $2,660. Thus,<br />

the CESR drops from .531 to 0 (1 – $2,660/$2,660).<br />

In the reinvestment case, the CE of the non-securitizing<br />

bank is $6,670 in Year 1, which comprises the PV of all current<br />

and future loan-related earnings of $6,397 and the PV<br />

of the current earnings in Year 1 reinvested for one year at a<br />

10% return rate ($273 = [$3,000 ¥ .1]/1.1). The reinvestment<br />

gains raise CESR to .550 compared with the distribution<br />

case. In Year 7, a CESR of .530 derives from the current<br />

earnings from loans ($3,000) and reinvestment<br />

($2,315) and from the PV of future earnings from loans<br />

($3,397) and reinvestment ($2,587), or .530 = 1 – ($3,000 +<br />

$2,315)/ ($3,000 + $2,315 + $3,397 + $2,587).<br />

Again, the values of the securitizing bank differ from<br />

those of the non-securitizing bank. When securitizing in<br />

Year 3, CESR drops to .147 because all future earnings<br />

from securitization ($6,397) are realized immediately along<br />

with the reinvestment gains ($630). Only the reinvestment<br />

generates future earnings ($1,212 = $1,333/1.1). Until Year<br />

7, CESR increases to .353 because gains from reinvestment<br />

rise but earnings from securitization remain constant. Note<br />

that CESR also reflects the investment horizon of the reinvestments,<br />

which is one year in our example. If this<br />

investment horizon was five years, it would correspond to<br />

the investment period of the (five-year) loans; therefore, the<br />

time horizon of both investments would be the same.<br />

The major advantage of CESR is that it creates more<br />

transparency and highlights, in a single metric, whether<br />

firms increase their current earnings at the expense of their<br />

long-term value. Comparing the distribution with the reinvestment<br />

case for the non-securitizing and securitizing<br />

bank reveals an additional strength of CESR: If the bank<br />

chooses to keep and reinvest its earnings from loans and<br />

securitization, the bank concurrently fosters its future earnings<br />

position, and CESR increases.<br />

Empirical Studies<br />

Aim<br />

First, we focus in our empirical studies on banking because<br />

of its extensive use of securitization. We determine whether<br />

banks provide sufficient information about the use of securitization<br />

to make their economic impact transparent. Then,<br />

124 / Journal of Marketing, May 2011<br />

we apply our reporting technique to the former market<br />

leader in mortgage lending and origination in the United<br />

States, Countrywide Financial Corporation, which made<br />

heavy use of securitization, and we show that CER and<br />

CESR could have provided much more transparency. Next,<br />

for nine securitizations of firms and institutions in diverse<br />

industries, we show that CESR can detect differences in the<br />

extent of the shift in earnings across time.<br />

Analysis of Securitization Transparency in<br />

Financial Statements<br />

We analyze 38 banks in the most important banking markets<br />

in the United States and Europe to determine how<br />

much information they provide about securitization in their<br />

financial statements (i.e., annual and quarterly reports). The<br />

sample includes two market leaders (measured by market<br />

capitalization) from ten countries and some randomly<br />

selected smaller banks and public institutions. These banks<br />

offer wide service portfolios to customers to originate sufficient<br />

loan volumes for securitization. As we show in Table<br />

3, nearly all banks are involved in securitization, but especially<br />

in the United States, few report the earnings that they<br />

realize separately. Instead, most banks follow the opportunity<br />

provided by current reporting standards and sum their<br />

“real interest earnings” plus “securitization earnings” as<br />

total interest income. This summary makes it difficult, if not<br />

impossible, to evaluate the extent of the shift toward shortterm<br />

profit realization inherent in securitization. In turn, it<br />

remains unclear what share of profit relates to the ongoing<br />

banking business and which part derives from the one-time<br />

effects of securitization.<br />

Description of Countrywide Financial Corporation<br />

Of this data set, we choose Countrywide for our detailed<br />

analysis because it fulfills our information requirements<br />

about securitization and other interest-bearing assets. In<br />

addition, Countrywide was the U.S. market leader in mortgage<br />

lending and origination between 2004 and 2007<br />

(Countrywide Financial Corp. 2008). Because of its heavy<br />

engagement in subprime mortgage–backed security transactions<br />

and the debt market turbulence of the financial crisis,<br />

Countrywide needed a rescue from Bank of America in<br />

February 2008 (Countrywide Financial Corp. 2008) and no<br />

longer publishes its own financial statements.<br />

Analysis of Countrywide Financial Corporation<br />

To apply CER and CESR to Countrywide, we make a few<br />

assumptions and calculations to project future earnings. The<br />

contracted loans generate net interest income and net loan<br />

servicing fees. We assume that loans are amortized linearly<br />

and that contracted earnings are realized over the loans’<br />

lifetimes. We also assume no further earnings from insurance<br />

premiums and other business after 2007 and adopt the<br />

distribution and reinvestment policy of Countrywide.<br />

In Table 4, we provide the resulting CE and CESR for<br />

1998–2007 for the actual business of Countrywide, which<br />

includes securitization (labeled “Securitization Case”). The<br />

increase in CE from $2.567 billion in 1998 to $20.151 billion<br />

in 2006 may be affiliated with higher loan volumes origi-


nated and sold, longer mortgage lifetimes within the servicing<br />

portfolio, or higher interest margins from the remaining<br />

business. Regardless of the cause, Countrywide created additional<br />

value for the future in 2006. The subsequent intense<br />

drop in earnings appeared in the 2007 CE of $9.647 billion,<br />

when the decreasing U.S. mortgage business reduced both<br />

total earnings and CE by approximately 50% in response to<br />

lower income levels and higher loan loss provisions.<br />

In contrast to the immense fluctuations in CE from 1998<br />

to 2007, CESR remained relatively stable, with values of<br />

.291 in 1998 and .372 in 2007. In other words, the earnings<br />

structure was not affected by the economic downturn, even<br />

when the business model suffered serious harm. Countrywide,<br />

with a CESR of .372 in 2007, still realized 62.8% of<br />

the earnings that it originated in the same year but only<br />

TAblE 3<br />

Information About Securitization from 38 banks<br />

General Securitization<br />

Interest Income<br />

from Other<br />

Securitization Interest-bearing<br />

bank<br />

largest National banks<br />

Country Information Volume Earnings Assets<br />

Erste Bank Austria √ √<br />

Raiffeisen International Bank Austria √ √<br />

BNP Paribas France √ √<br />

Société Générale France √ √<br />

Deutsche Bank Germany √ √<br />

Commerzbank Germany √ √<br />

Allied Irish Bank Ireland √ √<br />

Bank of Ireland Ireland √ √<br />

Unicredit Italy √<br />

Intesa San Paolo Italy √ √<br />

ING Group Netherlands √ √<br />

Fortis Netherlands √<br />

Santander Spain √ √<br />

BBVA Spain √ √<br />

Credit Suisse Switzerland √ √ √ √*<br />

UBS Switzerland √<br />

HSBC United Kingdom √ √<br />

Standard Chartered United Kingdom √ √<br />

JPMorgan Chase & Co. United States √ √ √ √*<br />

Wells Fargo<br />

Random Selection banks<br />

United States √ √ √ √*<br />

Bank Austria Austria √<br />

Comdirect Germany N.A.<br />

DAB Bank Germany N.A.<br />

Deutsche Postbank Germany √ √<br />

Hamburger Sparkasse Germany N.A.<br />

KfW Germany √ √<br />

Landesbank Hessen-Thüringen Germany √ √<br />

ABN Amro Netherlands √ √<br />

Rabobank Netherlands √<br />

Zürcher Kantonalbank Switzerland N.A.<br />

Barclays United Kingdom √ √<br />

HBOS United Kingdom √ √<br />

Lloyds TSB United Kingdom √ √<br />

Northern Rock United Kingdom √ √ (√)<br />

RBS United Kingdom √ √<br />

Bank of America United States √ √ √ √*<br />

Countrywide United States √ √ √ √*<br />

Wachovia United States √ √<br />

Notes: √ = Information is clearly available, √* = Information can be derived directly from the data, (√) = Information can be assumed from data,<br />

and N.A. = not available.<br />

37.2% in subsequent years. Because the average mortgage<br />

lifetime is 6.8 years, we believe it is fair to assert that Countrywide<br />

aimed to realize short-term profits.<br />

Because we want to demonstrate that CER and CESR<br />

could have depicted the realization of short-term profits at<br />

the expense of long-term value creation at Countrywide, we<br />

ran a counterfactual analysis (labeled “Non-securitization<br />

Case” in Table 4) in which we examined what would have<br />

happened had Countrywide not securitized at all. In this<br />

analysis, CE and CESR would have increased strongly. The<br />

most extreme difference would have occurred in 2002, when<br />

79.0% of Countrywide’s CE was immediately realized, but<br />

that value would have been just 32.7% in the securitization<br />

case. (For further details on these calculations, see Web<br />

Appendix A at http://www.marketingpower.com/ jmmay11.)<br />

Customer Equity Sustainability Ratio (CESR) / 125


TAblE 4<br />

Countrywide Earnings Structure<br />

A: Countrywide: Securitization Case (Actual business)<br />

1998 1999 2000 2001 2002 2003 2004 2005 2006 2007<br />

126 / Journal of Marketing, May 2011<br />

Originate and Distribute business<br />

Earnings on sale of mortgage loans<br />

and securities 699,433 557,743 611,092 1,498,812 3,377,065 5,541,539 4,386,536 4,300,579 4,897,771 2,220,164<br />

Earnings on sale of other loans and<br />

securities 623,531 406,458 398,544 103,178 94,153 345,897 455,546 561,201 784,076 214,559<br />

Total earnings, originate and distribute<br />

business 1,322,964 964,201 1,009,636 1,601,990 3,471,218 5,887,436 4,842,082 4,861,780 5,681,847 2,434,723<br />

buy and Hold business<br />

Net interest income 66,764 93,933 10,678 331,877 792,230 1,359,390 1,965,541 2,237,935 2,688,514 587,882<br />

Net loan servicing fees 241,817 551,723 584,024 –2,146 –865,752 –463,050 465,650 1,493,167 1,300,655 909,749<br />

Net insurance premiums earned 12,504 75,786 274,039 316,432 561,681 732,816 782,685 953,647 1,171,433 1,523,534<br />

Other 175,363 202,742 195,462 248,506 358,855 462,050 510,669 470,179 574,679 605,549<br />

Total earnings, buy and hold business 496,448 924,184 1,064,203 894,669 847,014 2,091,206 3,724,545 5,154,928 5,735,281 3,626,714<br />

Total earnings 1,819,412 1,888,385 2,073,839 2,496,659 4,318,232 7,978,642 8,566,627 10,016,708 11,417,128 6,061,437<br />

Customer equity 2,567,495 3,604,128 3,753,248 4,174,750 5,468,867 12,087,915 13,868,776 18,195,139 20,151,106 9,646,673<br />

Customer equity sustainability ratio .291 .476 .447 .402 .210 .340 .382 .449 .433 .372<br />

Securitization ratio 58.9% 65.3% 72.1% 95.5% 95.9% 86.1% 89.4% 82.5% 86.1% 90.4%<br />

Return on equity 15.3% 15.2% 11.6% 12.7% 18.2% 35.8% 23.9% 21.9% 19.7% –4.9%<br />

Nonprime loans as a percentage of total<br />

loan production 3.18% 2.69% 6.23% 4.50% 3.74% 4.56% 10.85% 8.94% 8.67% 4.09%<br />

Delinquent mortgage loans as a percentage<br />

of total loan production 3.91% 3.55% 11.30% 14.42% 14.41% 12.46% 11.29% 15.20% 19.03% 27.29%<br />

b: Countrywide: Non-Securitization Case (Counterfactual Analysis)<br />

1998 1999 2000 2001 2002 2003 2004 2005 2006 2007<br />

Total earnings 1,307,973 1,645,620 1,900,767 1,422,166 2,459,488 5,539,118 8,453,218 10,656,583 11,859,540 8,304,510<br />

Customer equity 2,567,495 4,166,711 4,639,131 5,339,600 7,517,683 16,039,134 20,618,857 25,532,604 27,518,455 17,264,104<br />

Customer equity sustainability ratio .491 .605 .590 .734 .673 .655 .590 .583 .569 .519<br />

Notes: Amounts listed are in thousands of U.S. dollars.


Application of the CESR to Industries Outside<br />

Banking<br />

Although securitization is most commonly used in banking,<br />

it enjoys great popularity in many other industries as well.<br />

Few of these firms provide sufficient information to illustrate<br />

the shift from long-term value creation to short-term<br />

profit realization either. In Table 5, we provide an overview<br />

of nine securitizations from a wide range of industries,<br />

including three state governments and several nonprofit<br />

organizations, which provide enough information to calcu-<br />

Reference<br />

Burke Sylva<br />

(1999)<br />

Campbell and<br />

Hashimoto<br />

(2003)<br />

EDF Group<br />

(2003)<br />

Everton Football<br />

Club (2002)<br />

Firm or<br />

Institution Transaction Type<br />

David Bowie In January 1997, David Bowie raised $55 million<br />

from the issue of ten-year asset-backed bonds,<br />

with collateral consisting of future royalties from<br />

25 albums that he recorded before 1990.<br />

Korean Air Lines Korean Air Lines securitized 2003 future<br />

receivables from the sale of its tickets on routes<br />

between Japan and Korea worth ¥27 billion<br />

($242 million) for a period of 36 months.<br />

EDF (Electricité<br />

de France)<br />

Everton FC<br />

(United Kingdom)<br />

In 1999, EDF transferred receivables of €1.1<br />

billion (US$1.2 billion in 1999) of employees’<br />

housing loans to the Electra mutual securitization<br />

fund. Since the end of 2000, EDF has<br />

transferred future trade receivables of energy<br />

supply contracts to a mutual securitization fund,<br />

reaching an amount of €2.1 billion ($1.8 billion in<br />

2001) by 2001.<br />

In 2002, Everton FC (UK soccer club) signed a<br />

securitization agreement serviced by future season<br />

ticket sales and match-day ticket sales. The<br />

£30 million ($45 million) loan must be repaid with<br />

the interest charge over a period of 25 years.<br />

FIFA (2006) FIFA FIFA, world soccer’s governing body, raised<br />

$420 million through a bond deal in 2001, using<br />

expected receipts of $536 million from the 2002<br />

World Cup in Japan and South Korea and the<br />

2006 World Cup in Germany as collateral.<br />

Armstrong (2009) Keele University<br />

(United Kingdom)<br />

Keele University (UK) securitized future student<br />

housing income worth £55.4 million (US$83.8<br />

million) to finance refurbishments of the student<br />

residences and repay existing debt.<br />

Kasprak (2002) State of Alaska Beginning in July 2001, Alaska used 40% of the<br />

tobacco settlement revenue to get an infusion of<br />

cash to build, rehabilitate, and remodel schools.<br />

The 2001 legislature passed in addition a law to<br />

securitize an additional 40% of the settlement<br />

revenue so that the transactions together securitized<br />

80% of the tobacco settlement revenue.<br />

Kasprak (2002) State of<br />

South Carolina<br />

TAblE 5<br />

Securitization in Industries Outside banking<br />

South Carolina passed legislation to securitize its<br />

MSA settlement dollars through 2018. The state<br />

established four funds to develop its water and<br />

wastewater infrastructure, compensate individuals<br />

for losses in tobacco production (25%), and<br />

fund a variety of health care programs and services<br />

(75%).<br />

Kasprak (2002) State of Wisconsin Wisconsin sold $1.54 billion of bonds backed by<br />

state appropriations to refinance debt issued by<br />

Badger Tobacco Asset Securitization Corp.<br />

late the effect of securitization on changes in CESR. (For<br />

more details on the calculations of securitization cases in<br />

industries outside banking, see Web Appendix B at http://<br />

www.marketingpower.com/jmmay11.)<br />

According to Table 5, the transaction volume ranged<br />

from several million to billions of U.S. dollars, and securitization<br />

appears in diverse settings. The most remarkable<br />

result is that the effects of securitization on changes in<br />

CESR differ across the nine securitizations and hardly correlate<br />

with transaction volume. For example, the largest<br />

securitization of $1.8 billion, for EDF Group, diminished<br />

CESR Without<br />

Securitization<br />

(1997):<br />

.911<br />

(2003):<br />

.833<br />

(1999):<br />

.908<br />

(2001):<br />

.882<br />

(2002):<br />

.989<br />

(2001):<br />

.943<br />

(2000):<br />

.938<br />

(2001):<br />

.906<br />

(2001):<br />

.906<br />

(2001):<br />

.915<br />

CESR with<br />

Securitization<br />

(1997):<br />

.230<br />

(2003):<br />

.825<br />

(1999):<br />

.905<br />

(2001):<br />

.876<br />

(2002):<br />

.958<br />

(2001):<br />

.876<br />

(2000):<br />

.891<br />

(2001):<br />

.181<br />

(2001):<br />

.104<br />

(2001):<br />

0<br />

Customer Equity Sustainability Ratio (CESR) / 127


CESR only from .882 to .876, and one of the smallest securitizations<br />

of $242 million, for Korean Air Lines, led to a<br />

comparable drop in the size of CESR. In contrast, the securitizations<br />

of David Bowie and the State of Wisconsin both<br />

prompted drops of CESR from approximately .900 to less<br />

than .250.<br />

In summary, although securitization is widespread in<br />

banking and prominent in the context of the financial crisis,<br />

it also is used commonly outside banking business. The<br />

diversity of firms and institutions indicate that securitization<br />

is applicable in many settings. Although we acknowledge<br />

that the lack of transparency requires strong assumptions<br />

about the missing information in some of the securitization<br />

cases in Table 5, CESR describes well how strongly the<br />

earnings shifted over time.<br />

Discussion and Conclusion<br />

Discussion of Customer Equity Sustainability<br />

Ratio<br />

Our empirical results illustrate that many banks and firms<br />

fail to provide sufficient transparency about their securitization<br />

activities, which makes it difficult, if not impossible, to<br />

evaluate which earnings come from ongoing business and<br />

which result from the one-time effects of securitization.<br />

Unfortunately, current reporting standards provide limited<br />

means to detect such behavior because firms are not<br />

required to report future earnings. Our empirical studies<br />

show that CER and the newly developed CESR provide<br />

more transparency and emphasize that firms that make<br />

extensive use of securitizations should increase their shortterm<br />

profits at the expense of the long-term value of their<br />

customer base. As forward-looking marketing metrics, CER<br />

and CESR add another perspective to backward-oriented<br />

accounting and reporting practices, which provide mainly<br />

historic information. We argue that marketing researchers<br />

have accumulated enough knowledge in the past decade to<br />

determine future values (e.g., by calculating CLV, CE). Ignorance<br />

of future earnings is no longer justified, and marketing<br />

academics should lead the field in the drive to consider future<br />

earnings in financial reporting. Such reporting techniques<br />

may also hasten marketing’s reentry into the boardroom.<br />

Comparison of Customer Equity Sustainability<br />

Ratio with Other Performance Ratios<br />

As a new key ratio, CESR requires comparisons with wellestablished<br />

performance ratios, such as ROE and return on<br />

risk-adjusted capital; efficiency ratios, such as the cost<br />

income ratio; and productivity ratios, including economic<br />

value added. 3 These ratios all use historic data and consider<br />

3Return on risk-adjusted capital = net income/allocated risk capital;<br />

this can be used to compare projects or investments under<br />

consideration on the basis of their implied risk profile. Cost<br />

income ratio = operating expenses/operating income—that is, the<br />

expenses needed to realize an income of $1. Economic value<br />

added = net operating profit – (capital ¥ cost of capital); economic<br />

value added was developed to measure the real surplus a company<br />

generates in a specific period from an appropriate return on the<br />

capital invested.<br />

128 / Journal of Marketing, May 2011<br />

only the current level of a firm’s profit or earnings, which<br />

means they neglect a possible shift from future (i.e., longterm)<br />

value creation to short-term profit realization. With<br />

securitization, the ratios exhibit better values in the respective<br />

year than they would otherwise because short-term profits<br />

decrease cost–income ratio and increase ROE, return on<br />

risk-adjusted capital, and economic value added. In contrast,<br />

securitization lowers CESR immediately and provides stakeholders<br />

with a reliable indicator that the firm has moved from<br />

long-term value creation to short-term profit realization.<br />

Another contribution of CESR becomes obvious when<br />

we compare it over time with the development of earnings,<br />

CE, ROE, and the securitization ratio (i.e., proportion of the<br />

securitized loan volume to total loan volume in a particular<br />

year) of Countrywide. Except for 2007, earnings and CE<br />

increased constantly, suggesting a successful business and<br />

value-creation process. In this empirical case, ROE would<br />

increase only after the securitization intensity had risen and<br />

then would return to its former levels. Furthermore, ROE<br />

could increase in response to other determinants, such as<br />

higher issued loan volumes, greater efficiency, and better<br />

equity structures. Therefore, a high ROE value can, but does<br />

not necessarily, substantiate heavy securitization engagement<br />

or short-term profit realizations.<br />

To determine the relationship between the securitization<br />

ratio and CESR, we use the results from our empirical study<br />

of Countrywide, which reveal an insignificant correlation<br />

of –.241 (p = .503, n = 10). This outcome indicates the<br />

securitization ratio cannot substitute for CESR. Although<br />

the changes in the securitization ratio point to a modification<br />

in the securitization strategy of Countrywide, it cannot<br />

indicate how much contracted business still gets realized in<br />

upcoming years, because this ratio usually measures the<br />

share of loans securitized in one particular year. Moreover,<br />

securitization ratios do not distinguish the case in which a<br />

firm distributes the earnings from securitization to its shareholders<br />

from one in which it reinvests these earnings and<br />

thus retains them in the firm.<br />

As an early warning system, CESR translates information<br />

into an earnings-related metric and includes earnings<br />

from nonsecuritization businesses. At the moment of securitization,<br />

CESR instantly declines, alerting stakeholders to<br />

the firm’s short-term profit realization. Furthermore, CESR<br />

enables stakeholders to recognize the firm’s distribution or<br />

reinvestment strategy, as we show in our example. Information<br />

about developments in the firm’s shift from long-term<br />

value creation to short-term profit realization and the comparisons<br />

with the values of other firms thus provide strong<br />

advantages. In particular, this ratio condenses important<br />

information without losing content and without limits on its<br />

applicability to a wide range of businesses. Legal and regulatory<br />

restrictions and the effort to keep competitive advantages<br />

have hindered firms from publishing a large pool of<br />

data on its future orientation, supporting the application of<br />

CER and CESR.<br />

Relationship Between Customer Equity and<br />

Customer Equity Sustainability Ratio<br />

To deliver the necessary level of transparency, firms should<br />

report both CE and CESR and grant stakeholders insights


into the value of their current customer base and the extent<br />

of their long-term value creation. As we highlight in Figure<br />

1, a high CE value indicates a strong customer base, which<br />

is always better than a weaker one. In contrast, CESR indicates<br />

whether the firm pursues long-term value creation or<br />

short-term profit realization. We recognize explicitly that<br />

CESR has no per se optimal value and leaves normative<br />

conclusions up to stakeholders. For example, if the firm is<br />

able to attract investors that are willing to pay a price above<br />

the fair value of the underlying loans, the sale is beneficial<br />

for the firm and its stakeholders even though securitization<br />

reduces CESR. The reverse holds if buyers are only willing<br />

to pay a price that is lower than the value for the seller. Still,<br />

if a firm with low values of CE and CESR reports high<br />

earnings for the current year, stakeholders should recognize<br />

its weak customer base and drop in future earnings if the<br />

firm cannot acquire enough valuable new customers or generate<br />

more business with its current customers.<br />

Even if sellers and buyers agree on the fair value of<br />

securities, extreme values of CESR could describe the best<br />

case for a firm under special circumstances. For example, a<br />

CESR of 0 can be the most desirable for a firm with liquidity<br />

needs that securitizes its entire future earnings to bridge<br />

the liquidity gap and likely prevent insolvency. For a project<br />

firm to develop and operate an office building with some<br />

lease contracts already signed at construction start, in the<br />

first years of development, CESR will be 1 because no<br />

earnings occur until the building is fully operational.<br />

Thus, there are no general critical values of CESR per<br />

se because CESR is business specific. However, stakeholders<br />

should be alerted if their firm’s CESR is fundamentally<br />

lower than the CESR in the firm’s peer group and be aware<br />

that their firm is focusing on short-term profit realization. In<br />

addition, if the firm’s CESR continuously declines over<br />

time, stakeholders should seriously consider a firm’s<br />

intended shift from long-term value creation to short-term<br />

profit realization. Moreover, it is necessary to exercise caution<br />

if huge increases in compensation or large dividends go<br />

along with strong declines in CESR.<br />

limitations and Further Research<br />

Our study is subject to some limitations that should prompt<br />

further research. First, the banking industry is complex,<br />

though we focus on simple standard loans in our example.<br />

Therefore, further research should identify the implied<br />

earnings structure of more complex products and business<br />

transactions so that CER and CESR can be implemented in<br />

diverse industries.<br />

Second, the aim of this article is to present CESR as a<br />

new metric and outline its benefits in several empirical studies.<br />

Further research should observe how regulatory authorities,<br />

particularly in the financial industry, can best establish<br />

obligations to publish CER and CESR. Securitizations can<br />

have a dramatic impact on variable payments in management<br />

compensation if these are linked to short-term profits.<br />

Further research could more strongly analyze how CER and<br />

CESR can help develop more appropriate compensation<br />

FIGURE 1<br />

Relationship between Customer Equity (CE) and<br />

Customer Equity Sustainability Ratio (CESR)<br />

CESR<br />

(Low)<br />

Short-term<br />

profit realization<br />

with a strong<br />

customer base<br />

Short-term<br />

profit realization<br />

with a weak<br />

customer base<br />

CE (High)<br />

Long-term<br />

value creation<br />

with a strong<br />

customer base<br />

Long-term<br />

value creation<br />

with a weak<br />

customer base<br />

CE (Low)<br />

CESR<br />

(High)<br />

plans and better disseminate information about managers’<br />

long-term value creation efforts.<br />

Third, the quality of CER and CESR crucially depends<br />

on good forecasts, particularly in industries with high<br />

shares of noncontractual business. Marketing researchers<br />

have developed powerful methods to accommodate the<br />

greater amount of uncertainty in such situations (e.g.,<br />

Reinartz and Kumar 2000), but the influence of such methods<br />

on current reporting standards is still limited. Determining<br />

the value of at least half of the nine securitizations displayed<br />

in Table 5 requires profound knowledge of the firm’s<br />

ability to further market the underlying products; however,<br />

currently, marketing scholars seem to play only a minor role<br />

in determining the value of these securitizations. Thus,<br />

future studies could outline in more detail how current marketing<br />

knowledge could be used to increase the quality of<br />

forecasts and determine the value of these securitizations<br />

more accurately.<br />

Fourth, current reporting standards already apply forwardlooking<br />

information and develop accounting forecasts based<br />

on future earnings, such as the fair values of goodwill and<br />

financial assets. External auditors assess uncertainty and reliability,<br />

limiting the space for managerial arbitrariness and<br />

opening the floor to our reporting technique. Therefore, further<br />

research should examine if CER and CESR require more<br />

detailed rules to reduce the risk of managerial arbitrariness.<br />

Fifth, securitization is likely to affect marketers and<br />

their customers, as we illustrate for a bank, though the<br />

effects hold equally for other firms. The lack of future earnings<br />

from the securitized loans of current customers obliges<br />

a bank to intensify its customer acquisition efforts and sell<br />

more loans to ensure that future earnings do not decrease, in<br />

particular if it distributes earnings to its shareholders. However,<br />

too much pressure on customer acquisition is likely to<br />

also increase the chances of acquiring unprofitable and<br />

high-risk customers (Cao and Gruca 2005; Reinartz,<br />

Thomas, and Kumar 2005), a factor that finance research<br />

Customer Equity Sustainability Ratio (CESR) / 129


also recognizes: “As balance sheets expand, new borrowers<br />

must be found. When all prime borrowers have a mortgage<br />

but balance sheets still need to expand, then banks have to<br />

lower their lending standards in order to lend to subprime<br />

borrowers. The seeds of the subsequent downturn in the<br />

credit cycle are thus sown” (Shin 2009, p. 310). Further<br />

research could provide stronger support for this speculative,<br />

albeit likely, statement.<br />

In summary, the lack of transparency in current financial<br />

reporting about the long-term effects of securitization is<br />

unacceptable. Customer equity reporting and the CESR<br />

REFERENCES<br />

Armstrong, Peter (2009), “Appraising Income Securitisation in the<br />

Public Sector: The Sale of Student Rents at Keele University,”<br />

International Journal of Critical Accounting, 1 (1/2), 3–57.<br />

Blattberg, Robert C. and John Deighton (1996), “Manage Marketing<br />

by the Customer Equity Test,” Harvard Business Review,<br />

74 (4), 136–44.<br />

Brinkworth, Stuart (2004), “Bondholders Cry Foul on Football<br />

Securitization,” International Financial Law Review, 23 (3),<br />

22–24.<br />

Burke Sylva, Jennifer (1999), “Bowie Bonds Sold for Far More<br />

than a Song: The Securitization of Intellectual Property as a<br />

Super-Charged Vehicle for High Technology Financing,” Santa<br />

Clara Computer & High Technology Law Journal, 15 (1),<br />

195–230.<br />

Calem, Paul S. and Michael LaCour-Little (2004), “Risk-Based<br />

Capital Requirements for Mortgage Loans,” Journal of Banking<br />

& <strong>Finance</strong>, 28 (3), 647–72.<br />

Campbell, Neil and Norifusa Hashimoto (2003), “Korean Air<br />

Lines Securitizes Future Yen Ticket Receivables,” International<br />

Financial Law Review, 22 (11), 16–17.<br />

Cao, Yong and Thomas S. Gruca (2005), “Reducing Adverse<br />

Selection Through Customer Relationship Management,” Journal<br />

of Marketing, 69 (October), 219–29.<br />

Countrywide Financial Corp. (2008), “Form 10-K,” annual report.<br />

Coval, Joshua D., Jakub Jurek, and Erik Stafford (2009), “The<br />

Economics of Structured <strong>Finance</strong>,” Journal of Economic Perspectives,<br />

23 (1), 3–25.<br />

Dechow, Patricia M., Linda A. Myers, and Catherine Shakespeare<br />

(2010), “Fair Value Accounting and Gains from Asset Securitizations:<br />

A Convenient Earnings Management Tool with Compensation<br />

Side-Benefits,” Journal of Accounting and Economics,<br />

49 (1), 2–25.<br />

——— and Catherine Shakespeare (2009), “Do Managers Time<br />

Securitization Transactions to Obtain Accounting Benefits?”<br />

Accounting Review, 84 (1), 99–132.<br />

Demyanyk, Yuliya and Otto Van Hemert (2009), “Understanding<br />

the Subprime Mortgage Crisis,” Review of Financial Studies,<br />

(published electronically May 4), [DOI: 10.1093/rfs/hhp033].<br />

Downey, Connor (1999), “Whole Business Securitization Comes<br />

of Age,” International Financial Law Review, 18 (September),<br />

8–12.<br />

EDF Group (2003), Annual Report 2003. Paris: EDF Group.<br />

Everton Football Club (2002), Annual Report & Statement of<br />

Accounts 2002. Liverpool, UK: Everton Football Club.<br />

Fabozzi, Frank J., Henry A. Davis, and Moorad Choudry (2006),<br />

Introduction to Structured <strong>Finance</strong>. Hoboken, NJ: John Wiley<br />

& Sons.<br />

FASB (1996), Statement of Financial Accounting Standards No.<br />

125. Norwalk, CT: Financial Accounting Standards Board.<br />

——— (2000), Statement of Financial Accounting Standards No.<br />

140. Norwalk, CT: Financial Accounting Standards Board.<br />

130 / Journal of Marketing, May 2011<br />

provide more transparency to stakeholders and demonstrate<br />

the financial stability of banks and firms. Perhaps CER and<br />

CESR could have helped avoid some of the challenges of<br />

the current financial crisis; more important, perhaps they<br />

can provide insights into how investors and firms can be<br />

supported in their attempts to prevent the next one. In any<br />

case, marketing should not leave the field to finance and<br />

accounting; instead, it should emphatically assert its claim<br />

and provide stakeholders with substantial information about<br />

the long-term value of the customer base.<br />

Franke, Günther and Jan P. Krahnen (2008), “The Future of Securitization,”<br />

paper presented at Brookings–Tokyo Club–Wharton<br />

Conference, Washington, DC (October 16).<br />

Greenbaum, Stuart I. and Anjan V. Thakor (1987), “Bank Funding<br />

Modes: Securitization Versus Deposits,” Journal of Banking &<br />

<strong>Finance</strong>, 11 (3), 379–401.<br />

Gregory, Michael (2002), “Broncos <strong>Finance</strong> Stadium via ABS,”<br />

Asset Securitization Report, 2 (7), 1–6.<br />

Gupta, Sunil, Donald R. Lehmann, and Jennifer Ames Stuart<br />

(2004), “Valuing Customers,” Journal of Marketing Research,<br />

41 (February), 7–18.<br />

International Federation of Association Football (2006), FIFA<br />

Financial Report 2006. Zurich: International Federation of<br />

Association Football.<br />

Joshi, Amit and Dominique M. Hanssens (2010), “The Direct and<br />

Indirect Effects of Advertising Spending on Firm Value,” Journal<br />

of Marketing, 74 (January), 20–33.<br />

Kasprak, John (2002), “Securitization of Tobacco Settlement<br />

Funds,” (accessed November 5, 2010), [available at http://<br />

www. cga.ct.gov/2002/olrdata/ph/rpt/2002-R-0736.htm].<br />

Kendall, Leon T. (1996), “Securitization: A New Era in American<br />

<strong>Finance</strong>,” in A Primer on Securitization, Leon T. Kendall and<br />

Michael J. Fishman, eds. Cambridge: MIT Press, 1–16.<br />

Ketkar, Suhas and Dilip Ratha (2008), “Development Financing<br />

During a Crisis: Securitization of Future Receivables,” working<br />

paper, The World Bank.<br />

Kothari, Vinod (2006), Securitization. Singapore: John Wiley &<br />

Sons.<br />

Luo, Dan, <strong>Dr</strong>agon Yongjun Tang, and Sarah Qian Wang (2009), “A<br />

Little Knowledge Is a Dangerous Thing: Data History, Model<br />

Uncertainty and CDO (Mis-) Pricing,” working paper, School<br />

of Economics and <strong>Finance</strong>, University of Hong Kong.<br />

McCreevy, Charlie (2008), “Commission Regulation (EC) No<br />

1126/2008,” Official Journal of the European Union, L 320,<br />

1–481.<br />

Reinartz, Werner J. and V. Kumar (2000), “On the <strong>Prof</strong>itability of<br />

Long-Life Customers in a Noncontractual Setting: An Empirical<br />

Investigation and Implications for Marketing,” Journal of<br />

Marketing, 64 (October), 17–35.<br />

———, Jacquelyn S. Thomas, and V. Kumar (2005), “Balancing<br />

Acquisition and Retention Resources to Maximize Customer<br />

<strong>Prof</strong>itability,” Journal of Marketing, 69 (January), 63–79.<br />

Rust, Roland T., Katherine N. Lemon, and Valarie A. Zeithaml<br />

(2004), “Return on Marketing: Using Customer Equity to<br />

Focus Marketing Strategy,” Journal of Marketing, 68 (January),<br />

109–127.<br />

———, Valarie A. Zeithaml, and Katherine N. Lemon (2000),<br />

<strong>Dr</strong>iving Customer Equity: How Customer Lifetime Value Is<br />

Reshaping Corporate Strategy. New York: The Free Press.<br />

Ryan, Stephen G. (2008), “Accounting in and for the Subprime<br />

Crisis,” Accounting Review, 83 (6), 1605–1638.


Santomero, Anthony M. and David F. Babbel (1997), Financial<br />

Markets, Instruments, and Institutions. Chicago: McGraw-Hill.<br />

Shin, Hyun Song (2009), “Securitisation and Financial Stability,”<br />

Economic Journal, 119 (March), 309–332.<br />

Srinivasan, Shuba and Dominique M. Hanssens (2009), “Marketing<br />

and Firm Value: Metrics, Methods, Findings, and Future<br />

Directions,” Journal of Marketing Research, 46 (June),<br />

293–312.<br />

Story, Louise, Landon Thomas Jr., and Nelson D. Schwartz<br />

(2010), “Wall Street Helped Mask Debt Fueling Europe’s Crisis,”<br />

The New York Times, (February 14), A1.<br />

Tuli, Kapil R., Sundar G. Bharadwaj, and Ajay K. Kohli (2010),<br />

“Ties That Bind: The Impact of Multiple Types of Ties with a<br />

Customer on Sales Growth and Sales Volatility,” Journal of<br />

Marketing Research, 47 (February), 36–50.<br />

Wiesel, Thorsten, Bernd Skiera, and Julián Villanueva (2008),<br />

“Customer Equity: An Integral Part of Financial Reporting,”<br />

Journal of Marketing, 72 (April), 1–14.<br />

Customer Equity Sustainability Ratio (CESR) / 131


Für weitere Informationen zum E-<strong>Finance</strong> <strong>Lab</strong> wenden Sie sich bitte an:<br />

<strong>Prof</strong>. <strong>Dr</strong>. <strong>Wolfgang</strong> <strong>König</strong><br />

Vorstandsvorsitzender des E-<strong>Finance</strong> <strong>Lab</strong>,<br />

Geschäftsführender Direktor des House of <strong>Finance</strong><br />

Grüneburgplatz 1<br />

60323 Frankfurt<br />

Telefon +49 (0)69 / 798 - 34001<br />

Fax +49 (0)69 / 798 - 33910<br />

E-Mail koenig@efinancelab.com<br />

<strong>Prof</strong>. <strong>Dr</strong>. Peter Gomber<br />

Stellvertretender Vorstandsvorsitzender des E-<strong>Finance</strong> <strong>Lab</strong><br />

Grüneburgplatz 1<br />

60323 Frankfurt<br />

Telefon +49 (0)69 / 798 - 34683<br />

Fax +49 (0)69 / 798 - 35007<br />

E-Mail gomber@wiwi.uni-frankfurt.de<br />

oder besuchen Sie unsere Webseite<br />

http://www.efinancelab.com

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!