• Ingen resultater fundet

Sparse linear manifolds relating shape to clinical outcome

N/A
N/A
Info
Hent
Protected

Academic year: 2022

Del "Sparse linear manifolds relating shape to clinical outcome "

Copied!
23
0
0

Indlæser.... (se fuldtekst nu)

Hele teksten

(1)

Sparse linear manifolds relating shape to clinical outcome

Professor , Ph.D. Rasmus Larsen

Hven , August 20st, 2009

DTU Informatics Technical University of Denmark

(2)

Purpose

We can extract measurements from the human body with a rapidly increasing spatial, temporal and spectral resolution using modern imaging devices. This is

particularly true in the field of biophotonics.

Typically we have an outcome (e.g. blood-glucose, psoriasis severity) that we want to predict based on a set of features (e.g. IR absorption spectra and derived features)

Having observed the outcome and features in a set of objects (a training set of data) we want to build a model that will allow us to predcit the outcome of unseen

objects

(3)

Model

Outcome: Y

Features: X = (X1, X2, … , Xp)

sampled spectrum

set of spectra in an image .

. .

Model: Y = f(X) + ε

Y

(4)

Two approaches

The linear model:

Global

Nearest Neighbour model:

Local

β ˆ

X

T

Y =

( ) ∑

=

) (

1

x N

i i

k

k y x

Y

(5)

Curse of dimensionality I

Consider inputs uniformly distributed over a p- dimensional hypercube [0,1]x[0,1]x…x[0,1]

2-dim hypercube:

For the red neighbourhood to cover a fraction r of the observation it should have side length s = r1/p

For r=1% we get for p=2: s = 0.1, for p=10: s=0.63

(6)

Curse of dimensionality II

For practical size problems locality in high dimensional spaces does not exist

The majority of observations lie near the edges of the training sample, in the 10 dimensional hypercube, only 1% of the observations lie in a central hypercube of sidelength 0.63 – we must extrapolate our fits

In high dimensions the linear model is popular!

(7)

Linear Regression

Training set

(8)

Linear Regression – matrix-vector notation

The predictor Xβ belongs to the column-space of X

(9)

Linear regression - geometrically

Choose β such that the residual is orthogonal to X, i.e.

(10)

Linear regression – correlated inputs

XTX/N is the ML estimator for the covariance matrix of the inputs Consider 3 inputs X1 , X2 , X3 with covariance

=

1 0

0

0 1

99 . 0

0 99

. 0 1

S

=

1 0

0

0 25

. 50 75

. 49

0 75

. 49 25

. 50 S 1

The parameters of the correlated inputs have high variance and high correlation

(11)

Linear regression – regularization

(12)

Ridge regression- geometrically

(13)

Ridge regression – geometrically II

β1 β2

β1 β2

(14)

Correllated inputs again

3 inputs X1 , X2 , X3 with covariance

Y = X1 + X2 + X3 + ε, ε in N(0,1)

N=100 , in 1000 trials

Ordinary LS

=

1 0

0

0 1

99 . 0

0 99

. 0 1

S

( )

=

08 . 1 56

. 56

. 0

56 . 0 55

55

56 . 0 55

55 100

β 1 Cov

β = [-0.01 0.97 1.03 1.00

(15)

Correllated inputs again – ridge regression

(λ, RSS)

( )

=

02 . 1 03

. 05

. 0

03 . 0 3

. 4 8

. 3

05 . 0 8

. 3 4

. 4 100

β 1

Ridge (λ=2.4) Cov

β = [-0.00 0.99 0.98 0.98

(16)

We want

Prediction accuracy

Easy Intepretation (simple model) We tried

Regularization (ridge regression) And got

Prediction accuracy

(17)

Prediction accuracy and easy interpretation

many β’s will tend to be 0

Regularization and subset selection

β1 β2

(18)

LASSO Model Selection

0 5 10 15 20 25 30

-1 -0.5

0 0.5

1 1.5

2

step β

(19)

LASSO

Prediction accuracy ☺ Easy interpretation ☺ Computations ☺

p<N

Tend to select one of a group of correlated inputs

(20)

LARS-EN – elastic net

Prediction accuracy Easy interpretation ☺ Computations

Handles p>N

Tend to select groups of correlated inputs ☺

(21)

LARS-EN – elastic net

Ridge to OLS

LASSO problem remains!

(22)

Handling CoD

Regularization

Variable selection Subspace projection

(23)

Principal Components

By rotating the coordinate system, the axes point in directions of maximum variance

22/34

XL S =

The new axes are in the loading matrix Coordinates of data on new axes

are in the scores matrix

data matrix

Referencer

RELATEREDE DOKUMENTER

To put the data-driven SPCA method into perspective, tests for each clinical outcome variable were also investigated through a direct analysis of the original variables and by using

We have shown how to use efficient first-order convex optimization techniques in a multiple description framework in order to form sparse descriptions, which satisfies a set

Knowledge of the course and outcome of major affective illness has clinical as well as theoretical implications. In understanding the pathophysiology of the major affective

I want to identify a task for the writer that will transform the project (or the part of it we’re working on in the moment), without predetermining the outcome, in a process that

In comparison, other member states with a tradition of constitutional democracy, such as Germany, do not experience CJEU interference as threatening towards national

We use a set of Stochastic Frontier Analysis (SFA) models to estimate cost functions that allow us to examine the impact of regional-level economic and institutional factors on

The recordTarget records the patient whose health information (in the context of this IG, patient responses to a set of questions (or patient reported outcome measure) asked

To investigate the optimal number of data points for training, we will test the accuracy and training time as a function of the number of training data points.. The parameter