29.11.2012 Views

2nd USENIX Conference on Web Application Development ...

2nd USENIX Conference on Web Application Development ...

2nd USENIX Conference on Web Application Development ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Latency (µs)<br />

1e+06<br />

100000<br />

10000<br />

1000<br />

TTPCase<br />

XenCase<br />

BaseCase<br />

100<br />

0.1 1 10 100<br />

Invocati<strong>on</strong> Size (KB)<br />

Latency (ms)<br />

100000<br />

10000<br />

1000<br />

100<br />

10<br />

1<br />

TTPCase<br />

XenCase<br />

BaseCase<br />

0.1<br />

1 10 100 1000<br />

Number of Invocati<strong>on</strong>s<br />

34 <strong>Web</strong>Apps ’11: <str<strong>on</strong>g>2nd</str<strong>on</strong>g> <str<strong>on</strong>g>USENIX</str<strong>on</strong>g> <str<strong>on</strong>g>C<strong>on</strong>ference</str<strong>on</strong>g> <strong>on</strong> <strong>Web</strong> Applicati<strong>on</strong> <strong>Development</strong> <str<strong>on</strong>g>USENIX</str<strong>on</strong>g> Associati<strong>on</strong><br />

Latency (ms)<br />

1000<br />

100<br />

10<br />

1<br />

TTPCase<br />

XenCase<br />

BaseCase<br />

10 100<br />

Number of Ads<br />

Figure 3: Invocati<strong>on</strong> costs: (left) measurement preserver; latency vs. payload size, (middle) stock preserver; latency vs. number of<br />

ticker events, (right) targeted ads preservers; latency vs number of ads.<br />

Stock Preserver:To reflect a typical trading scenario with<br />

hundreds of ticker events per sec<strong>on</strong>d, we plotted latency<br />

versus numbers of back-to-back invocati<strong>on</strong>s (stock ticker<br />

events) in Figure 3(middle). As expected, the TTP case<br />

requiring network access is substantially slower as compared<br />

to the Xen case and the base case. Comparing the<br />

Xen case and the base case for a sequence of 100 backto-back<br />

invocati<strong>on</strong>s, the latencies are 43.4ms and 9.05ms<br />

respectively; though the overhead due to Xen is substantial,<br />

it is still a significant improvement over the TTP<br />

case, which requires 12 sec<strong>on</strong>ds. Regarding the traffic,<br />

in the TTP case, the traffic generated for a sequence of<br />

5, 10, 50, 100, 500 and 1000 invocati<strong>on</strong>s are respectively<br />

2.8, 4.8, 20.2, 39.4, 194, 387 KB respectively. Thus, the<br />

bandwidth required for say, 500 events per sec<strong>on</strong>d, is<br />

about 1.6 MB/s per user per stock symbol (and a single<br />

user typically trades multiple symbols). These results reflect<br />

that server site colocati<strong>on</strong> can offer significant benefit<br />

in terms of network bandwidth.<br />

Targeted Ads Preserver: We present two results for the<br />

targeted ads scenario. Figure 3(right) shows the latency<br />

per invocati<strong>on</strong> for the preserver to serve targeted ads using<br />

the ChooseAd interface (we do not show results for<br />

the GetInterestVector interface due to lack of space). This<br />

graph shows that the Xen preserver has nearly the same<br />

overhead as the Base case preserver <strong>on</strong>ce the number<br />

of ads out of which the preserver selects <strong>on</strong>e exceeds<br />

5. This is mainly because the payload size dominates<br />

the transfer time; the c<strong>on</strong>text switch overhead is minimal.<br />

In the TTP case, the latency is clearly impacted by<br />

the wide-area network delay. The traffic generated in the<br />

TTP case was measured to be 2.6, 8.3, 15.8, 29.2 KB for<br />

10, 50, 100, 200 ads respectively. These reflect the bandwidth<br />

savings of colocati<strong>on</strong>; since this bandwidth is incurred<br />

for every website the user visits, this could be significant.<br />

Our technical report [16] has performance results<br />

<strong>on</strong> two other operati<strong>on</strong>s: allowing a user to update her web<br />

history and the data aggregati<strong>on</strong> transformati<strong>on</strong>.<br />

6.3 Security Analysis<br />

We now discuss the desirable security properties of our<br />

preserver, determine the TCB for these to hold, and argue<br />

that our TCB is a significant improvement over the status<br />

quo. Our security goal is that any data access or preserver<br />

transfer obeys the user-sancti<strong>on</strong>ed interface and policies.<br />

Depending <strong>on</strong> the deployment, the adversary either has<br />

c<strong>on</strong>trol of the service software stack (above the VMM<br />

layer) or has physical access to the machine hosting the<br />

preserver; the first applies to VMMs, TPMs, and the sec<strong>on</strong>d<br />

to TTPs, client-side preservers, secure co-processors.<br />

In the TTP/client case, we rely <strong>on</strong> physical isolati<strong>on</strong> and<br />

the preserver implementati<strong>on</strong>’s correctness; the rest of this<br />

secti<strong>on</strong> is c<strong>on</strong>cerned with the security goal for colocati<strong>on</strong>.<br />

Colocati<strong>on</strong> has <strong>on</strong>e basic caveat: the service can launch<br />

an availability attack by refusing to service requests using<br />

its host hub or by preventing a user from updating her<br />

preserver. We assume that the trusted module provides the<br />

four security properties detailed in Secti<strong>on</strong> 3: isolati<strong>on</strong>,<br />

anti-replay protecti<strong>on</strong>, random number generati<strong>on</strong>, and remote<br />

attestati<strong>on</strong>. We also assume that any side-channels<br />

(such as memory page caching, CPU usage) have limited<br />

bandwidth (a subject of <strong>on</strong>going research [30]). We first<br />

analyze the installati<strong>on</strong> protocol and then examine the invocati<strong>on</strong><br />

protocol.<br />

Installati<strong>on</strong> Protocol: The installati<strong>on</strong> protocol is<br />

somewhat complex since it involves a Diffie-Hellman key<br />

exchange al<strong>on</strong>g with remote attestati<strong>on</strong>. In order to argue<br />

its correctness, we now present an informal argument (our<br />

technical report [16] has a formal argument in LS2 [10],<br />

a formal specificati<strong>on</strong> language). Our informal argument<br />

is based <strong>on</strong> three salient features of the installati<strong>on</strong> protocol<br />

(Secti<strong>on</strong> 4.1). First, the attestati<strong>on</strong> guarantees that<br />

the public key is generated by the base layer. Sec<strong>on</strong>d, the<br />

Diffie-Hellman-based installati<strong>on</strong> protocol ensures c<strong>on</strong>fidentiality;<br />

<strong>on</strong>ly the party that generated the public key itself<br />

can decrypt the preserver. From the first observati<strong>on</strong>,<br />

this party can <strong>on</strong>ly be the base layer. This relies <strong>on</strong> the<br />

random number generati<strong>on</strong> ability of the trusted module<br />

to ensure that the generated public key is truly random.<br />

Third, since attestati<strong>on</strong> verifies the purported input and<br />

output to the code which generated the public key, manin-the-middle<br />

attacks are ruled out; an adversary cannot<br />

tamper with the attestati<strong>on</strong> without rendering it invalid.<br />

These three properties together ensure that the preserver

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!