Slide 1

Slide 1 text

REGRESSION VS. APPROXIMATION What it means for uncertainty quantification PAUL CONSTANTINE Ben L. Fryrear Assistant Professor Applied Mathematics & Statistics Colorado School of Mines activesubspaces.org! @DrPaulynomial! SLIDES AVAILABLE UPON REQUEST DISCLAIMER: These slides are meant to complement the oral presentation. Use out of context at your own risk.

Slide 2

Slide 2 text

data model noise parameter error uncertainty overfit TRIGGER WARNING!

Slide 3

Slide 3 text

Selected regression / approximation / UQ literature My personal bibliography Ghanem and Spanos, Stochastic Finite Elements (Springer, 1991) Xiu and Karniadakis, The Wiener-Askey polynomial chaos (SISC, 2002) Nobile, Tempone, and Webster, A sparse grid stochastic collocation method (SINUM, 2008) Gautschi, Orthogonal Polynomials (Oxford UP, 2004) Koehler and Owen, Computer experiments (Handbook of Statistics, 1996) Jones, A taxonomy of global optimization methods based on response surfaces (JGO, 2001) Cook, Regression Graphics (Wiley, 1998)

Slide 4

Slide 4 text

No content

Slide 5

Slide 5 text

No content

Slide 6

Slide 6 text

Regression Approximation x y x y

Slide 7

Slide 7 text

Regression { xi, yi } ⇡ ( x, y ) i.i.d. samples from unknown GIVEN GOAL statistically characterize y | x e.g., E[ y | x ] , Var[ y | x ] ˆ ✓ = argmin ✓ X i yi p ( xi, ✓ ) 2 y = p ( x, ✓ ) + " MODEL (e.g., polynomials) E[ y | x ] modeled r.v., zero-mean, independent of x x y FIT (e.g., max likelihood)

Slide 8

Slide 8 text

Regression x y { xi, yi } ⇡ ( x, y ) i.i.d. samples from unknown GIVEN GOAL statistically characterize y | x e.g., E[ y | x ] , Var[ y | x ] ˆ ✓ = argmin ✓ X i yi p ( xi, ✓ ) 2 y = p ( x, ✓ ) + " E[ y | x ] modeled r.v., zero-mean, independent of x PREDICT E[ y | x ⇤ ] ⇡ p ( x ⇤ , ˆ ✓ ) = ˆ p ( x ⇤) MODEL (e.g., polynomials) FIT (e.g., max likelihood)

Slide 9

Slide 9 text

Var[ y | x ⇤ ] ⇡ “formula” E[ y | x ⇤ ] ⇡ p ( x ⇤ , ˆ ✓ ) = ˆ p ( x ⇤) ˆ ✓ = argmin ✓ X i yi p ( xi, ✓ ) 2 y = p ( x, ✓ ) + " Regression x y { xi, yi } ⇡ ( x, y ) i.i.d. samples from unknown GIVEN GOAL statistically characterize y | x e.g., E[ y | x ] , Var[ y | x ] E[ y | x ] modeled r.v., zero-mean, independent of x FIT (e.g., max likelihood) PREDICT QUANTIFY UNCERTAINTY MODEL (e.g., polynomials)

Slide 10

Slide 10 text

k p⇤ f k = e⇤(n) Approximation continuous functions polynomials of degree n p⇤ = argmin p 2 Pn k p f k Does a unique, best approximation exist? polynomials How does the best error behave? continuous function Can we construct an approximation? Algorithm: Given f , compute ˆ p k ˆ p f k  C e⇤(n) And analyze its error?

Slide 11

Slide 11 text

kˆ p fk Approximation a function GIVEN GOAL f ( x ) ⇡ ( x ) find such that the error is small a known density ˆ p ( x ) x y

Slide 12

Slide 12 text

kˆ p fk Approximation x y a function GIVEN GOAL f ( x ) ⇡ ( x ) find such that the error is small a known density CONSTRUCTION ˆ p ( x ) choose xi

Slide 13

Slide 13 text

kˆ p fk Approximation x y a function GIVEN GOAL f ( x ) ⇡ ( x ) find such that the error is small a known density CONSTRUCTION ˆ p ( x ) choose xi compute yi = f ( xi)

Slide 14

Slide 14 text

kˆ p fk Approximation x y GIVEN GOAL ⇡ ( x ) find such that the error is small a known density CONSTRUCTION choose ˆ p ( x ) xi compute yi = f ( xi) fit ˆ p = argmin p 2 Pn X i yi p ( xi) 2 a function f ( x )

Slide 15

Slide 15 text

Regression Approximation x y x y The story of the data and fitted curve is different. But does it matter? YES ˆ p = argmin p 2 Pn X i yi p ( xi) 2 ˆ ✓ = argmin ✓ X i yi p ( xi, ✓ ) 2

Slide 16

Slide 16 text

REGRESSION VS. APPROXIMATION What is error?

Slide 17

Slide 17 text

Regression Approximation x y x y Approximation error | ˆ p ( x ) f ( x ) | sup x | ˆ p ( x ) f ( x ) | ✓Z | ˆ p ( x ) f ( x ) |2 dx ◆1/2 Error norms Confidence interval ˆ p ( x ) ± 2 b se[ y | x ] plug-in estimate of standard error

Slide 18

Slide 18 text

REGRESSION VS. APPROXIMATION What is convergence?

Slide 19

Slide 19 text

Regression Approximation x y x y ˆ ✓ ! ✓ ˆ p ( x ) ! p ( x ) “true” parameters As data increases, root-n consistency ˆ p ( x ) f ( x ) ! 0 As the approximation class grows Convergence rate depends on •  high order derivatives •  size of region of analyticity •  Chebyshev coefficients •  … f ( x )

Slide 20

Slide 20 text

Regression Approximation x y x y Which one is a computer simulation?

Slide 21

Slide 21 text

Sacks et al. (1989) What is error in a computer simulation? Kennedy and O’Hagan (2001)

Slide 22

Slide 22 text

What is error in a computer simulation? Oberkampf et al. (2002) von Neumann and Goldstine Bulletin of the AMS (1947) [h/t Joe Grcar]

Slide 23

Slide 23 text

“This analysis of the sources of errors should be objective and strict inasmuch as completeness is concerned, but when it comes to the defining, classifying, and separating of the sources, a certain subjectiveness and arbitrariness is unavoidable. With these reservations, the following enumeration and classification of sources of errors seems to be adequate and reasonable.” Mathematical model Observations and parameters Finitistic approximations Round-off The von Neumann and Goldstine Catechism

Slide 24

Slide 24 text

“This analysis of the sources of errors should be objective and strict inasmuch as completeness is concerned, but when it comes to the defining, classifying, and separating of the sources, a certain subjectiveness and arbitrariness is unavoidable. With these reservations, the following enumeration and classification of sources of errors seems to be adequate and reasonable.” Mathematical model Observations and parameters Finitistic approximations Round-off The von Neumann and Goldstine Catechism NOTES How well math model approximates reality Model-form error

Slide 25

Slide 25 text

“This analysis of the sources of errors should be objective and strict inasmuch as completeness is concerned, but when it comes to the defining, classifying, and separating of the sources, a certain subjectiveness and arbitrariness is unavoidable. With these reservations, the following enumeration and classification of sources of errors seems to be adequate and reasonable.” Mathematical model Observations and parameters Finitistic approximations Round-off The von Neumann and Goldstine Catechism NOTES Forward and inverse UQ Most of the UQ methods literature

Slide 26

Slide 26 text

“This analysis of the sources of errors should be objective and strict inasmuch as completeness is concerned, but when it comes to the defining, classifying, and separating of the sources, a certain subjectiveness and arbitrariness is unavoidable. With these reservations, the following enumeration and classification of sources of errors seems to be adequate and reasonable.” Mathematical model Observations and parameters Finitistic approximations Round-off The von Neumann and Goldstine Catechism NOTES Asymptotics from classical numerical analysis Deterministic numerical noise “Computational noise in deterministic simulations is as ill-defined a concept as can be found in scientific computing.” Moré and Wild (2011)

Slide 27

Slide 27 text

I think UQ is more approximation than regression, because computer models are deterministic. But computational noise is really annoying, if you take it seriously. LOTS of fundamental research opportunities for applying statistical methods to noise-less data---i.e., the approximation setting. Summary thoughts

Slide 28

Slide 28 text

No content

Slide 29

Slide 29 text

Active subspaces The null space of the matrix DIMENSION REDUCTION reveals directions in the space of along which constant. f( x ) x Z rf( x ) rf( x )T ⇢( x ) d x Constantine, Dow, and Wang (2014)

Slide 30

Slide 30 text

What is sufficient dimension reduction for a deterministic function of several variables?

Slide 31

Slide 31 text

www.siam.org/meetings/dr17/ http://www.siam.org/journals/juq/juq_special.php

Slide 32

Slide 32 text

PAUL CONSTANTINE Ben L. Fryrear Assistant Professor Colorado School of Mines activesubspaces.org! @DrPaulynomial! QUESTIONS? Active Subspaces SIAM (2015) How many samples do you need? What is the error? Which variables are the most important?