jagomart
digital resources
picture1_Statistical Methods Pdf 87125 | Notes Msm


 154x       Filetype PDF       File size 0.44 MB       Source: www.statslab.cam.ac.uk


File: Statistical Methods Pdf 87125 | Notes Msm
modern statistical methods rajen d shah r shah statslab cam ac uk course webpage http www statslab cam ac uk rds37 modern stat methods html in this course we will ...

icon picture PDF Filetype PDF | Posted on 14 Sep 2022 | 3 years ago
Partial capture of text on file.
                             Modern Statistical Methods
                           Rajen D. Shah            r.shah@statslab.cam.ac.uk
                                             Course webpage:
                    http://www.statslab.cam.ac.uk/ rds37/modern_stat_methods.html
                                                    ~
                In this course we will study a selection of important modern statistical methods. This
             selection is heavily biased towards my own interests, but I hope it will nevertheless give you
             a flavour of some of the most important recent methodological developments in statistics.
                Over the last 25 years, the sorts of datasets that statisticians have been challenged to
             study have changed greatly. Where in the past, we were used to datasets with many obser-
             vations with a few carefully chosen variables, we are now seeing datasets where the number
             of variables can run into the thousands and greatly exceed the number of observations. For
             example, with microarray data, we typically have gene expression values measured for sev-
             eral thousands of genes, but only for a few hundred tissue samples. The classical statistical
             methods are often simply not applicable in these “high-dimensional” situations.
                The course is divided into 4 chapters (of unequal size). Our first chapter will start by
             introducing ridge regression, a simple generalisation of ordinary least squares. Our study
             of this will lead us to some beautiful connections with functional analysis and ultimately
             one of the most successful and flexible classes of learning algorithms: kernel machines.
                The second chapter concerns the Lasso and its extensions. The Lasso has been at the
             centre of much of the developments that have occurred in high-dimensional statistics, and
             will allow us to perform regression in the seemingly hopeless situation when the number
             of parameters we are trying to estimate is larger than the number of observations.
                In the third chapter we will study graphical modelling and provide an introduction to
             the exciting field of causal inference. Where the previous chapters consider methods for
             relating a particular response to a large collection of (explanatory) variables, graphical
             modelling will give us a way of understanding relationships between the variables them-
             selves. Ultimately we would like to infer causal relationships between variables based on
             (observational) data. This may seem like a fundamentally impossible task, yet we will show
             howbydeveloping the graphical modelling framework further, we can begin to answer such
             causal questions.
                Statistics is not only about developing methods that can predict well in the presence
             of noise, but also about assessing the uncertainty in our predictions and estimates. In
             the final chapter we will tackle the problem of how to handle performing thousands of
             hypothesis tests at the same time and more generally the task of quantifying uncertainty
             in high-dimensional settings.
                Before we begin the course proper, we will briefly review two key classical statistical
             methods: ordinary least squares and maximum likelihood estimation. This will help to set
             the scene and provide a warm-up for the modern methods to come later.
                                                    i
                Classical statistics
                Ordinary least squares
                Imagine data are available in the form of observations (Y ,x ) ∈ R × Rp, i = 1,...,n, and
                                                                                 i  i
                the aim is to infer a simple regression function relating the average value of a response, Yi,
                and a collection of predictors or variables, xi. This is an example of regression analysis,
                one of the most important tasks in statistics.
                    Alinear model for the data assumes that it is generated according to
                                                           Y =Xβ0+ε,                                          (0.0.1)
                                n                                         n×p
                where Y ∈ R is the vector of responses; X ∈ R                  is the predictor matrix (or design
                matrix) with ith row xT; ε ∈ Rn represents random error; and β0 ∈ Rp is the unknown
                                           i
                vector of coefficients.
                    Provided p ≪ n, a sensible way to estimate β is by ordinary least squares (OLS). This
                                        OLS
                                       ˆ
                yields an estimator β        with
                                           OLS                         2       T    −1   T
                                         ˆ
                                         β     := argmin kY −Xβk2 = (X X) X Y,                                (0.0.2)
                                                    β∈Rp
                provided X has full column rank.
                    Under the assumptions that (i) E(ε ) = 0 and (ii) Var(ε) = σ2I, we have that:
                                                            i
                                OLS            T   −1   T      0           0
                          0 2 ˆ
                    • E      (β     ) = E{(X X) X (Xβ +ε)}=β .
                         β ,σ
                                  OLS         T    −1  T              T    −1     2   T    −1
                            0  2 ˆ
                    • Varβ ,σ (β      ) = (X X) X Var(ε)X(X X)                =σ (X X) .
                    The Gauss–Markov theorem states that OLS is the best linear unbiased estimator in
                                                         ˜                          ˜
                our setting: for any other estimator β that is linear in Y (so β = AY for some fixed matrix
                A), we have
                                                                              OLS
                                                        0  2 ˜          0 2 ˆ
                                                   Varβ ,σ (β) − Varβ ,σ (β       )
                is positive semi-definite.
                Maximum likelihood estimation
                The method of least squares is just one way to construct as estimator. A more general
                technique is that of maximum likelihood estimation. Here given data y ∈ Rn that we take
                as a realisation of a random variable Y , we specify its density f(y;θ) up to some unknown
                                                    d
                vector of parameters θ ∈ Θ ⊆ R , where Θ is the parameter space. The likelihood function
                is a function of θ for each fixed y given by
                                                   L(θ) := L(θ;y) = c(y)f(y;θ),
                where c(y) is an arbitrary constant of proportionality. The maximum likelihood estimate
                of θ maximises the likelihood, or equivalently it maximises the log-likelihood
                                             ℓ(θ) := ℓ(θ;y) = logf(y;θ) + log(c(y)).
                                                                  ii
                  Avery useful quantity in the context of maximum likelihood estimation is the Fisher
              information matrix with jkth (1 ≤ j,k ≤ d) entry
                                                          ∂2         
                                            ijk(θ) := −Eθ  ∂θj∂θkℓ(θ) .
              It can be thought of as a measure of how hard it is to estimate θ when it is the true
                                                                            ˜
              parameter value. The Cram´er–Rao lower bound states that if θ is an unbiased estimator
              of θ, then under regularity conditions,
                                                        ˜    −1
                                                  Varθ(θ)−i (θ)
              is positive semi-definite.
                  Aremarkable fact about maximum likelihood estimators (MLEs) is that (under quite
              general conditions) they are asymptotically normally distributed, asymptotically unbiased
              and asymptotically achieve the Cram´er–Rao lower bound.
                  AssumethattheFisherinformationmatrixwhentherearenobservations,i(n)(θ)(where
              we have made the dependence on n explicit) satisfies i(n)(θ)/n → I(θ) for some positive
              definite matrix I. Then denoting the maximum likelihood estimator of θ when there are
                                 (n)
                                ˆ
              n observations by θ   , under regularity conditions, as the number of observations n → ∞
              we have                       √
                                                 (n)     d        −1
                                                ˆ
                                              n(θ   −θ)→Nd(0,I (θ)).
                  Returning to our linear model, if we assume in addition that ε ∼ N(0,σ2), then the
                                                                                i
              log-likelihood for (β,σ2) is
                                                                  n
                                          2      n      2     1 X          T  2
                                     ℓ(β,σ ) = − log(σ )−           (y −x β) .
                                                 2           2σ2      i    i
                                                                 i=1
              We see that the maximum likelihood estimate of β and OLS coincide. It is easy to check
              that                                                     
                                          i(β,σ2) =   σ−2XTX        0     .
                                                          0      nσ−4/2
                                                                            √ ˆ                  2  T    −1
              ThegeneraltheoryforMLEswouldsuggestthatapproximately           n(β−β)∼Np(0,nσ (X X) );
              in fact it is straight-forward to show that this distributional result is exact.
                                                         iii
                Contents
                1 Kernel machines                                                                               1
                    1.1  Ridge regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .      1
                         1.1.1   The singular value decomposition and principal components analysis              3
                    1.2  v-fold cross-validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     4
                    1.3  The kernel trick . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .      5
                    1.4  Kernels . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     7
                         1.4.1   Examples of kernels . . . . . . . . . . . . . . . . . . . . . . . . . . .       8
                         1.4.2   Reproducing kernel Hilbert spaces . . . . . . . . . . . . . . . . . . .         9
                         1.4.3   The representer theorem . . . . . . . . . . . . . . . . . . . . . . . .       12
                    1.5  Kernel ridge regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     13
                    1.6  Other kernel machines . . . . . . . . . . . . . . . . . . . . . . . . . . . . .       16
                         1.6.1   The support vector machine . . . . . . . . . . . . . . . . . . . . . .        16
                         1.6.2   Logistic regression    . . . . . . . . . . . . . . . . . . . . . . . . . . .  18
                    1.7  Large-scale kernel machines      . . . . . . . . . . . . . . . . . . . . . . . . . .  19
                2 The Lasso and beyond                                                                         21
                    2.1  Model selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     21
                    2.2  The Lasso estimator      . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .  22
                         2.2.1   Prediction error of the Lasso with no assumptions on the design . .           23
                         2.2.2   Basic concentration inequalities . . . . . . . . . . . . . . . . . . . .      23
                         2.2.3   Some facts from optimisation theory and convex analysis . . . . . .           27
                         2.2.4   Lasso solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     30
                         2.2.5   Variable selection . . . . . . . . . . . . . . . . . . . . . . . . . . . .    30
                         2.2.6   Prediction and estimation . . . . . . . . . . . . . . . . . . . . . . .       31
                         2.2.7   Computation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .       35
                    2.3  Extensions of the Lasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . .     36
                         2.3.1   Structural penalties . . . . . . . . . . . . . . . . . . . . . . . . . . .    36
                         2.3.2   Reducing the bias of the Lasso . . . . . . . . . . . . . . . . . . . . .      37
                3 Graphical modelling and causal inference                                                     38
                    3.1  Graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .    38
                    3.2  Conditional independence graphs . . . . . . . . . . . . . . . . . . . . . . .         40
                    3.3  Gaussian graphical models . . . . . . . . . . . . . . . . . . . . . . . . . . .       41
                                                                iv
The words contained in this file might help you see if this file matches what you are looking for:

...Modern statistical methods rajen d shah r statslab cam ac uk course webpage http www rds stat html in this we will study a selection of important is heavily biased towards my own interests but i hope it nevertheless give you avour some the most recent methodological developments statistics over last years sorts datasets that statisticians have been challenged to changed greatly where past were used with many obser vations few carefully chosen variables are now seeing number can run into thousands and exceed observations for example microarray data typically gene expression values measured sev eral genes only hundred tissue samples classical often simply not applicable these high dimensional situations divided chapters unequal size our rst chapter start by introducing ridge regression simple generalisation ordinary least squares lead us beautiful connections functional analysis ultimately one successful exible classes learning algorithms kernel machines second concerns lasso its extensi...

no reviews yet
Please Login to review.