28.01.2015 Views

Slides in PDF - of Marcus Hutter

Slides in PDF - of Marcus Hutter

Slides in PDF - of Marcus Hutter

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Marcus</strong> <strong>Hutter</strong> - 37 - Universal Induction & Intelligence<br />

Generalization: Cont<strong>in</strong>uous Classes M<br />

In statistical parameter estimation one <strong>of</strong>ten has a cont<strong>in</strong>uous<br />

hypothesis class (e.g. a Bernoulli(θ) process with unknown θ ∈ [0, 1]).<br />

∫<br />

∫<br />

M := {ν θ : θ ∈ IR d }, ξ(x) := dθ w(θ) ν θ (x), dθ w(θ) = 1<br />

IR d<br />

Under weak regularity conditions [CB90,H’03]:<br />

IR d<br />

Theorem: D n (µ||ξ) ≤ ln w(µ) −1 + d 2 ln n 2π + O(1)<br />

where O(1) depends on the local curvature (parametric complexity) <strong>of</strong><br />

ln ν θ , and is <strong>in</strong>dependent n for many reasonable classes, <strong>in</strong>clud<strong>in</strong>g all<br />

stationary (k th -order) f<strong>in</strong>ite-state Markov processes (k = 0 is i.i.d.).<br />

D n ∝ log(n) = o(n) still implies excellent prediction and decision for<br />

most n.<br />

[RH’07]

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!