目录:Matrix Differential Calculus with Applications in Statistics and Econometrics,3rd_[Magnus2019]...

目录:Matrix Differential Calculus with Applications in Statistics and Econometrics,3rd_[Magnus2019]

Title   -16
Contents    -14
Preface -6
Part One — Matrices 1
1 Basic properties of vectors and matrices  3
    1.1 Introduction    3
    1.2 Sets    3
    1.3 Matrices: addition and multiplication   4
    1.4 The transpose of a matrix   6
    1.5 Square matrices 6
    1.6 Linear forms and quadratic forms    7
    1.7 The rank of a matrix    9
    1.8 The inverse 10
    1.9 The determinant 10
    1.10 The trace  11
    1.11 Partitioned matrices   12
    1.12 Complex matrices   14
    1.13 Eigenvalues and eigenvectors   14
    1.14 Schur’s decomposition theorem  17
    1.15 The Jordan decomposition   18
    1.16 The singular-value decomposition   20
    1.17 Further results concerning eigenvalues 20
    1.18 Positive (semi)definite matrices   23
    1.19 Three further results for positive definite matrices   25
    1.20 A useful result    26
    1.21 Symmetric matrix functions 27
    Miscellaneous exercises 28
    Bibliographical notes   30
2 Kronecker products, vec operator, and Moore-Penrose inverse   31
    2.1 Introduction    31
    2.2 The Kronecker product   31
    2.3 Eigenvalues of a Kronecker product  33
    2.4 The vec operator    34
    2.5 The Moore-Penrose (MP) inverse  36
    2.6 Existence and uniqueness of the MP inverse  37
    2.7 Some properties of the MP inverse   38
    2.8 Further properties  39
    2.9 The solution of linear equation systems 41
    Miscellaneous exercises 43
    Bibliographical notes   45
3 Miscellaneous matrix results  47
    3.1 Introduction    47
    3.2 The adjoint matrix  47
    3.3 Proof of Theorem 3.1    49
    3.4 Bordered determinants   51
    3.5 The matrix equation AX = 0  51
    3.6 The Hadamard product    52
    3.7 The commutation matrix K mn 54
    3.8 The duplication matrix D n  56
    3.9 Relationship between D n+1 and D n , I  58
    3.10 Relationship between D n+1 and D n , II    59
    3.11 Conditions for a quadratic form to be positive (negative) subject to linear constraints    60
    12 Necessary and sufficient conditions for r(A : B) = r(A) + r(B)63
    13 The bordered Gramian matrix  65
    14 The equations X 1 A + X 2 B ′ = G 1 ,X 1 B = G 2 67
    Miscellaneous exercises 69
    Bibliographical notes   70
Part Two — Differentials: the theory
4 Mathematical preliminaries    73
    4.1 Introduction    73
    4.2 Interior points and accumulation points 73
    4.3 Open and closed sets    75
    4.4 The Bolzano-Weierstrass theorem 77
    4.5 Functions   78
    4.6 The limit of a function 79
    4.7 Continuous functions and compactness    80
    4.8 Convex sets 81
    4.9 Convex and concave functions    83
    Bibliographical notes   86
5 Differentials and differentiability   87
    5.1 Introduction    87
    5.2 Continuity  88
    5.3 Differentiability and linear approximation  90
    5.4 The differential of a vector function   91
    5.5 Uniqueness of the differential  93
    5.6 Continuity of differentiable functions  94
    5.7 Partial derivatives 95
    5.8 The first identification theorem    96
    5.9 Existence of the differential, I    97
    5.10 Existence of the differential, II  99
    5.11 Continuous differentiability   100
    5.12 The chain rule 100
    5.13 Cauchy invariance  102
    5.14 The mean-value theorem for real-valued functions   103
    5.15 Differentiable matrix functions    104
    5.16 Some remarks on notation   106
    5.17 Complex differentiation    108
    Miscellaneous exercises 110
    Bibliographical notes   110
6 The second differential   111
    6.1 Introduction    111
    6.2 Second-order partial derivatives    111
    6.3 The Hessian matrix  112
    6.4 Twice differentiability and second-order approximation, I   113
    6.5 Definition of twice differentiability   114
    6.6 The second differential 115
    6.7 Symmetry of the Hessian matrix  117
    6.8 The second identification theorem   119
    6.9 Twice differentiability and second-order approximation, II  119
    6.10 Chain rule for Hessian matrices    121
    6.11 The analog for second differentials    123
    6.12 Taylor’s theorem for real-valued functions 124
    6.13 Higher-order differentials 125
    6.14 Real analytic functions    125
    6.15 Twice differentiable matrix functions  126
    Bibliographical notes   127
7 Static optimization   129
    7.1 Introduction    129
    7.2 Unconstrained optimization  130
    7.3 The existence of absolute extrema   131
    7.4 Necessary conditions for a local minimum    132
    7.5 Sufficient conditions for a local minimum: first-derivative test    134
    7.6 Sufficient conditions for a local minimum: second-derivative test   136
    7.7 Characterization of differentiable convex functions 138
    7.8 Characterization of twice differentiable convex functions   141
    7.9 Sufficient conditions for an absolute minimum   142
    7.10 Monotonic transformations  143
    7.11 Optimization subject to constraints    144
    7.12 Necessary conditions for a local minimum under constraints 145
    7.13 Sufficient conditions for a local minimum under constraints    149
    7.14 Sufficient conditions for an absolute minimum under constraints    154
    7.15 A note on constraints in matrix form   155
    7.16 Economic interpretation of Lagrange multipliers    155
    Appendix: the implicit function theorem 157
    Bibliographical notes   159
Part Three — Differentials: the practice    161
8 Some important differentials  163
    8.1 Introduction    163
    8.2 Fundamental rules of differential calculus  163
    8.3 The differential of a determinant   165
    8.4 The differential of an inverse  168
    8.5 Differential of the Moore-Penrose inverse   169
    8.6 The differential of the adjoint matrix  172
    8.7 On differentiating eigenvalues and eigenvectors 174
    8.8 The continuity of eigenprojections  176
    8.9 The differential of eigenvalues and eigenvectors: symmetric case    180
    8.10 Two alternative expressions for dλ 183
    8.11 Second differential of the eigenvalue function 185
    Miscellaneous exercises 186
    Bibliographical notes   189
9 First-order differentials and Jacobian matrices   191
    9.1 Introduction    191
    9.2 Classification  192
    9.3 Derisatives 192
    9.4 Derivatives 194
    9.5 Identification of Jacobian matrices 196
    9.6 The first identification table  197
    9.7 Partitioning of the derivative  197
    9.8 Scalar functions of a scalar    198
    9.9 Scalar functions of a vector    198
    9.10 Scalar functions of a matrix, I: trace 199
    9.11 Scalar functions of a matrix, II: determinant  201
    9.12 Scalar functions of a matrix, III: eigenvalue  202
    9.13 Two examples of vector functions   203
    9.14 Matrix functions   204
    9.15 Kronecker products 206
    9.16 Some other problems    208
    9.17 Jacobians of transformations   209
    Bibliographical notes   210
10 Second-order differentials and Hessian matrices  211
    10.1 Introduction   211
    10.2 The second identification table    211
    10.3 Linear and quadratic forms 212
    10.4 A useful theorem   213
    10.5 The determinant function   214
    10.6 The eigenvalue function    215
    10.7 Other examples 215
    10.8 Composite functions    217
    10.9 The eigenvector function   218
    10.10 Hessian of matrix functions, I    219
    10.11 Hessian of matrix functions, II   219
    Miscellaneous exercises 220
Part Four — Inequalities    223
11 Inequalities 225
    11.1 Introduction   225
    11.2 The Cauchy-Schwarz inequality  226
    11.3 Matrix analogs of the Cauchy-Schwarz inequality    227
    11.4 The theorem of the arithmetic and geometric means  228
    11.5 The Rayleigh quotient  230
    11.6 Concavity of λ 1 and convexity of λ n  232
    11.7 Variational description of eigenvalues 232
    11.8 Fischer’s min-max theorem  234
    11.9 Monotonicity of the eigenvalues    236
    11.10 The Poincar´ e separation theorem 236
    11.11 Two corollaries of Poincar´ e’s theorem   237
    11.12 Further consequences of the Poincar´ e theorem    238
    11.13 Multiplicative version    239
    11.14 The maximum of a bilinear form    241
    11.15 Hadamard’s inequality 242
    11.16 An interlude: Karamata’s inequality   242
    11.17 Karamata’s inequality and eigenvalues 244
    11.18 An inequality concerning positive semidefinite matrices   245
    11.19 A representation theorem for (Papi) 1/p   246
    11.20 A representation theorem for (trA p ) 1/p 247
    11.21 Hölder’s inequality   248
    11.22 Concavity of log|A|   250
    11.23 Minkowski’s inequality    251
    11.24 Quasilinear representation of |A| 1/n 253
    11.25 Minkowski’s determinant theorem   255
    11.26 Weighted means of order p 256
    11.27 Schl¨ omilch’s inequality 258
    11.28 Curvature properties of M p (x,a) 259
    11.29 Least squares 260
    11.30 Generalized least squares 261
    11.31 Restricted least squares  262
    11.32 Restricted least squares: matrix version  264
    Miscellaneous exercises 265
    Bibliographical notes   269
Part Five — The linear model    271
12 Statistical preliminaries    273
    12.1 Introduction   273
    12.2 The cumulative distribution function   273
    12.3 The joint density function 274
    12.4 Expectations   274
    12.5 Variance and covariance    275
    12.6 Independence of two random variables   277
    12.7 Independence of n random variables 279
    12.8 Sampling   279
    12.9 The one-dimensional normal distribution    279
    12.10 The multivariate normal distribution  280
    12.11 Estimation    282
    Miscellaneous exercises 282
    Bibliographical notes   283
13 The linear regression model  285
    13.1 Introduction   285
    13.2 Affine minimum-trace unbiased estimation   286
    13.3 The Gauss-Markov theorem   287
    13.4 The method of least squares    290
    13.5 Aitken’s theorem   291
    13.6 Multicollinearity  293
    13.7 Estimable functions    295
    13.8 Linear constraints: the case M(R ′ ) ⊂ M(X ′ ) 296
    13.9 Linear constraints: the general case   300
    13.10 Linear constraints: the case M(R ′ ) ∩ M(X ′ ) = {0}  302
    13.11 A singular variance matrix: the case M(X) ⊂ M(V ) 304
    13.12 A singular variance matrix: the case r(X'V+ X) = r(X) 305
    13.13 A singular variance matrix: the general case, I   307
    13.14 Explicit and implicit linear constraints  307
    13.15 The general linear model, I   310
    13.16 A singular variance matrix: the general case, II  311
    13.17 The general linear model, II  314
    13.18 Generalized least squares 315
    13.19 Restricted least squares  316
    Miscellaneous exercises 318
    Bibliographical notes   319
14 Further topics in the linear model   321
    14.1 Introduction   321
    14.2 Best quadratic unbiased estimation of σ 2  322
    14.3 The best quadratic and positive unbiased estimator of σ 2  322
    14.4 The best quadratic unbiased estimator of σ 2   324
    14.5 Best quadratic invariant estimation of σ 2 326
    14.6 The best quadratic and positive invariant estimator of σ 2 327
    14.7 The best quadratic invariant estimator of σ 2  329
    14.8 Best quadratic unbiased estimation: multivariate normal case   330
    14.9 Bounds for the bias of the least-squares estimator of σ 2 , I  332
    14.10 Bounds for the bias of the least-squares estimator of σ 2 , II    333
    14.11 The prediction of disturbances    335
    14.12 Best linear unbiased predictors with scalar variance matrix   336
    14.13 Best linear unbiased predictors with fixed variance matrix, I 338
    14.14 Best linear unbiased predictors with fixed variance matrix, II    340
    14.15 Local sensitivity of the posterior mean   341
    14.16 Local sensitivity of the posterior precision  342
    Bibliographical notes   344
Part Six — Applications to maximum likelihood estimation    345
15 Maximum likelihood estimation    347
    15.1 Introduction   347
    15.2 The method of maximum likelihood (ML)  347
    15.3 ML estimation of the multivariate normal distribution  348
    15.4 Symmetry: implicit versus explicit treatment   350
    15.5 The treatment of positive definiteness 351
    15.6 The information matrix 352
    15.7 ML estimation of the multivariate normal distribution:distinct means   354
    15.8 The multivariate linear regression model   354
    15.9 The errors-in-variables model  357
    15.10 The nonlinear regression model with normal errors 359
    15.11 Special case: functional independence of mean and variance parameters 361
    15.12 Generalization of Theorem 15.6    362
    Miscellaneous exercises 364
    Bibliographical notes   365
16 Simultaneous equations   367
    16.1 Introduction   367
    16.2 The simultaneous equations model   367
    16.3 The identification problem 369
    16.4 Identification with linear constraints on B and Γ only 371
    16.5 Identification with linear constraints on B, Γ, and Σ  371
    16.6 Nonlinear constraints  373
    16.7 FIML: the information matrix (general case)    374
    16.8 FIML: asymptotic variance matrix (special case)    376
    16.9 LIML: first-order conditions   378
    16.10 LIML: information matrix  381
    16.11 LIML: asymptotic variance matrix  383
    Bibliographical notes   388
17 Topics in psychometrics  389
    17.1 Introduction   389
    17.2 Population principal components    390
    17.3 Optimality of principal components 391
    17.4 A related result   392
    17.5 Sample principal components    393
    17.6 Optimality of sample principal components  395
    17.7 One-mode component analysis    395
    17.8 One-mode component analysis and sample principal components    398
    17.9 Two-mode component analysis    399
    17.10 Multimode component analysis  400
    17.11 Factor analysis   404
    17.12 A zigzag routine  407
    17.13 A Newton-Raphson routine  408
    17.14 Kaiser’s varimax method   412
    17.15 Canonical correlations and variates in the population 414
    17.16 Correspondence analysis   417
    17.17 Linear discriminant analysis  418
    Bibliographical notes   419
Part Seven — Summary    421
18 Matrix calculus: the essentials  423
    18.1 Introduction   423
    18.2 Differentials  424
    18.3 Vector calculus    426
    18.4 Optimization   429
    18.5 Least squares  431
    18.6 Matrix calculus    432
    18.7 Interlude on linear and quadratic forms    434
    18.8 The second differential    434
    18.9 Chain rule for second differentials    436
    18.10 Four examples 438
    18.11 The Kronecker product and vec operator    439
    18.12 Identification    441
    18.13 The commutation matrix    442
    18.14 From second differential to Hessian   443
    18.15 Symmetry and the duplication matrix   444
    18.16 Maximum likelihood    445
    Further reading 448
Bibliography    449
Index of symbols    467
Subject index   471

转载于:https://www.cnblogs.com/ourweiguan/p/10986690.html

Contents Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii Part One — Matrices 1 Basic properties of vectors and matrices3 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . .3 2Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .3 3Matrices: addition and multiplication . . . . . . . . . . . . . . .4 4The transpose of a matrix . . . . . . . . . . . . . . . . . . . . .6 5Square matrices . . . . . . . . . . . . . . . . . . . . . . . . . . .6 6Linear forms and quadratic forms . . . . . . . . . . . . . . . . .7 7The rank of a matrix . . . . . . . . . . . . . . . . . . . . . . . .8 8The inverse . . . . . . . . . . . . . . . . . . . . . . . . . . . . .9 9The determinant . . . . . . . . . . . . . . . . . . . . . . . . . . 10 10 The trace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 11 Partitioned matrices . . . . . . . . . . . . . . . . . . . . . . . . 11 12 Complex matrices . . . . . . . . . . . . . . . . . . . . . . . . . 13 13 Eigenvalues and eigenvectors . . . . . . . . . . . . . . . . . . . 14 14 Schur’s decomposition theorem . . . . . . . . . . . . . . . . . . 17 15 The Jordan decomposition . . . . . . . . . . . . . . . . . . . . . 18 16 The singular-value decomposition . . . . . . . . . . . . . . . . . 19 17 Further results concerning eigenvalues . . . . . . . . . . . . . . 20 18 Positive (semi)definite matrices . . . . . . . . . . . . . . . . . . 23 19 Three further results for positive definite matrices . . . . . . . 25 20 A useful result . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 27 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 2 Kronecker products, the vec operator and the Moore-Penrose inverse 31 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 2The Kronecker product . . . . . . . . . . . . . . . . . . . . . . 31 3Eigenvalues of a Kronecker product . . . . . . . . . . . . . . . . 33 4The vec operator . . . . . . . . . . . . . . . . . . . . . . . . . . 34 5The Moore-Penrose (MP) inverse . . . . . . . . . . . . . . . . . 36 6Existence and uniqueness of the MP inverse . . . . . . . . . . . 37viContents 7Some properties of the MP inverse . . . . . . . . . . . . . . . . 38 8Further properties . . . . . . . . . . . . . . . . . . . . . . . . . 39 9The solution of linear equation systems . . . . . . . . . . . . . 41 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 43 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 3 Miscellaneous matrix results47 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 2The adjoint matrix . . . . . . . . . . . . . . . . . . . . . . . . . 47 3Proof of Theorem 1 . . . . . . . . . . . . . . . . . . . . . . . . . 49 4Bordered determinants . . . . . . . . . . . . . . . . . . . . . . . 51 5The matrix equation AX = 0 . . . . . . . . . . . . . . . . . . . 51 6The Hadamard product . . . . . . . . . . . . . . . . . . . . . . 53 7The commutation matrix Kmn. . . . . . . . . . . . . . . . . . 54 8The duplication matrix Dn. . . . . . . . . . . . . . . . . . . . 56 9Relationship between Dn+1and Dn, I . . . . . . . . . . . . . . 58 10 Relationship between Dn+1and Dn, II . . . . . . . . . . . . . . 60 11 Conditions for a quadratic form to be positive (negative) sub- ject to linear constraints . . . . . . . . . . . . . . . . . . . . . . 61 12 Necessary and sufficient conditions for r(A : B) = r(A) + r(B)64 13 The bordered Gramian matrix . . . . . . . . . . . . . . . . . . 66 14 The equations X1A + X2B′= G1,X1B = G2. . . . . . . . . . 68 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 71 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 Part Two — Differentials: the theory 4 Mathematical preliminaries75 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75 2Interior points and accumulation points . . . . . . . . . . . . . 75 3Open and closed sets . . . . . . . . . . . . . . . . . . . . . . . . 76 4The Bolzano-Weierstrass theorem . . . . . . . . . . . . . . . . . 79 5Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 6The limit of a function . . . . . . . . . . . . . . . . . . . . . . . 81 7Continuous functions and compactness . . . . . . . . . . . . . . 82 8Convex sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83 9Convex and concave functions . . . . . . . . . . . . . . . . . . . 85 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 5 Differentials and differentiability89 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 2Continuity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 3Differentiability and linear approximation . . . . . . . . . . . . 91 4The differential of a vector function . . . . . . . . . . . . . . . . 93 5Uniqueness of the differential . . . . . . . . . . . . . . . . . . . 95 6Continuity of differentiable functions . . . . . . . . . . . . . . . 96 7Partial derivatives . . . . . . . . . . . . . . . . . . . . . . . . . 97Contentsvii 8The first identification theorem . . . . . . . . . . . . . . . . . . 98 9Existence of the differential, I . . . . . . . . . . . . . . . . . . . 99 10 Existence of the differential, II . . . . . . . . . . . . . . . . . . 101 11 Continuous differentiability . . . . . . . . . . . . . . . . . . . . 103 12 The chain rule . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 13 Cauchy invariance . . . . . . . . . . . . . . . . . . . . . . . . . 105 14 The mean-value theorem for real-valued functions . . . . . . . . 106 15 Matrix functions . . . . . . . . . . . . . . . . . . . . . . . . . . 107 16 Some remarks on notation . . . . . . . . . . . . . . . . . . . . . 109 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 110 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 111 6 The second differential113 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113 2Second-order partial derivatives . . . . . . . . . . . . . . . . . . 113 3The Hessian matrix . . . . . . . . . . . . . . . . . . . . . . . . . 114 4Twice differentiability and second-order approximation, I . . . 115 5Definition of twice differentiability . . . . . . . . . . . . . . . . 116 6The second differential . . . . . . . . . . . . . . . . . . . . . . . 118 7(Column) symmetry of the Hessian matrix . . . . . . . . . . . . 120 8The second identification theorem . . . . . . . . . . . . . . . . 122 9Twice differentiability and second-order approximation, II . . . 123 10 Chain rule for Hessian matrices . . . . . . . . . . . . . . . . . . 125 11 The analogue for second differentials . . . . . . . . . . . . . . . 126 12 Taylor’s theorem for real-valued functions . . . . . . . . . . . . 128 13 Higher-order differentials . . . . . . . . . . . . . . . . . . . . . . 129 14 Matrix functions . . . . . . . . . . . . . . . . . . . . . . . . . . 129 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 131 7 Static optimization133 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133 2Unconstrained optimization . . . . . . . . . . . . . . . . . . . . 134 3The existence of absolute extrema . . . . . . . . . . . . . . . . 135 4Necessary conditions for a local minimum . . . . . . . . . . . . 137 5Sufficient conditions for a local minimum: first-derivative test . 138 6Sufficient conditions for a local minimum: second-derivative test140 7Characterization of differentiable convex functions . . . . . . . 142 8Characterization of twice differentiable convex functions . . . . 145 9Sufficient conditions for an absolute minimum . . . . . . . . . . 147 10 Monotonic transformations . . . . . . . . . . . . . . . . . . . . 147 11 Optimization subject to constraints . . . . . . . . . . . . . . . . 148 12 Necessary conditions for a local minimum under constraints . . 149 13 Sufficient conditions for a local minimum under constraints . . 154 14 Sufficient conditions for an absolute minimum under constraints158 15 A note on constraints in matrix form . . . . . . . . . . . . . . . 159 16 Economic interpretation of Lagrange multipliers . . . . . . . . . 160 Appendix: the implicit function theorem . . . . . . . . . . . . . . . . 162viiiContents Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 163 Part Three — Differentials: the practice 8 Some important differentials167 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167 2Fundamental rules of differential calculus . . . . . . . . . . . . 167 3The differential of a determinant . . . . . . . . . . . . . . . . . 169 4The differential of an inverse . . . . . . . . . . . . . . . . . . . 171 5Differential of the Moore-Penrose inverse . . . . . . . . . . . . . 172 6The differential of the adjoint matrix . . . . . . . . . . . . . . . 175 7On differentiating eigenvalues and eigenvectors . . . . . . . . . 177 8The differential of eigenvalues and eigenvectors: symmetric case 179 9The differential of eigenvalues and eigenvectors: complex case . 182 10 Two alternative expressions for dλ . . . . . . . . . . . . . . . . 185 11 Second differential of the eigenvalue function . . . . . . . . . . 188 12 Multiple eigenvalues . . . . . . . . . . . . . . . . . . . . . . . . 189 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 189 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 192 9 First-order differentials and Jacobian matrices193 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193 2Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193 3Bad notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 194 4Good notation . . . . . . . . . . . . . . . . . . . . . . . . . . . 196 5Identification of Jacobian matrices . . . . . . . . . . . . . . . . 198 6The first identification table . . . . . . . . . . . . . . . . . . . . 198 7Partitioning of the derivative . . . . . . . . . . . . . . . . . . . 199 8Scalar functions of a vector . . . . . . . . . . . . . . . . . . . . 200 9Scalar functions of a matrix, I: trace . . . . . . . . . . . . . . . 200 10 Scalar functions of a matrix, II: determinant . . . . . . . . . . . 202 11 Scalar functions of a matrix, III: eigenvalue . . . . . . . . . . . 204 12 Two examples of vector functions . . . . . . . . . . . . . . . . . 204 13 Matrix functions . . . . . . . . . . . . . . . . . . . . . . . . . . 205 14 Kronecker products . . . . . . . . . . . . . . . . . . . . . . . . . 208 15 Some other problems . . . . . . . . . . . . . . . . . . . . . . . . 210 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 211 10 Second-order differentials and Hessian matrices213 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213 2The Hessian matrix of a matrix function . . . . . . . . . . . . . 213 3Identification of Hessian matrices . . . . . . . . . . . . . . . . . 214 4The second identification table . . . . . . . . . . . . . . . . . . 215 5An explicit formula for the Hessian matrix . . . . . . . . . . . . 217 6Scalar functions . . . . . . . . . . . . . . . . . . . . . . . . . . . 217 7Vector functions . . . . . . . . . . . . . . . . . . . . . . . . . . 219 8Matrix functions, I . . . . . . . . . . . . . . . . . . . . . . . . . 220Contentsix 9Matrix functions, II . . . . . . . . . . . . . . . . . . . . . . . . 221 Part Four — Inequalities 11 Inequalities225 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 225 2The Cauchy-Schwarz inequality . . . . . . . . . . . . . . . . . . 225 3Matrix analogues of the Cauchy-Schwarz inequality . . . . . . . 227 4The theorem of the arithmetic and geometric means . . . . . . 228 5The Rayleigh quotient . . . . . . . . . . . . . . . . . . . . . . . 230 6Concavity of λ1, convexity of λn. . . . . . . . . . . . . . . . . 231 7Variational description of eigenvalues . . . . . . . . . . . . . . . 232 8Fischer’s min-max theorem . . . . . . . . . . . . . . . . . . . . 233 9Monotonicity of the eigenvalues . . . . . . . . . . . . . . . . . . 235 10 The Poincar´e separation theorem . . . . . . . . . . . . . . . . . 236 11 Two corollaries of Poincar´e’s theorem . . . . . . . . . . . . . . 237 12 Further consequences of the Poincar´e theorem . . . . . . . . . . 238 13 Multiplicative version . . . . . . . . . . . . . . . . . . . . . . . 239 14 The maximum of a bilinear form . . . . . . . . . . . . . . . . . 241 15 Hadamard’s inequality . . . . . . . . . . . . . . . . . . . . . . . 242 16 An interlude: Karamata’s inequality . . . . . . . . . . . . . . . 243 17 Karamata’s inequality applied to eigenvalues . . . . . . . . . . 245 18 An inequality concerning positive semidefinite matrices . . . . . 245 19 A representation theorem for (Pap i)1/p. . . . . . . . . . . . . 246 20 A representation theorem for (trAp)1/p. . . . . . . . . . . . . . 248 21 H¨older’s inequality . . . . . . . . . . . . . . . . . . . . . . . . . 249 22 Concavity of log|A| . . . . . . . . . . . . . . . . . . . . . . . . . 250 23 Minkowski’s inequality . . . . . . . . . . . . . . . . . . . . . . . 252 24 Quasilinear representation of |A|1/n. . . . . . . . . . . . . . . . 254 25 Minkowski’s determinant theorem . . . . . . . . . . . . . . . . . 256 26 Weighted means of order p . . . . . . . . . . . . . . . . . . . . . 256 27 Schl¨omilch’s inequality . . . . . . . . . . . . . . . . . . . . . . . 259 28 Curvature properties of Mp(x,a) . . . . . . . . . . . . . . . . . 260 29 Least squares . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261 30 Generalized least squares . . . . . . . . . . . . . . . . . . . . . 263 31 Restricted least squares . . . . . . . . . . . . . . . . . . . . . . 263 32 Restricted least squares: matrix version . . . . . . . . . . . . . 265 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 266 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 270 Part Five — The linear model 12 Statistical preliminaries275 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 275 2The cumulative distribution function . . . . . . . . . . . . . . . 275 3The joint density function . . . . . . . . . . . . . . . . . . . . . 276 4Expectations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 276xContents 5Variance and covariance . . . . . . . . . . . . . . . . . . . . . . 277 6Independence of two random variables . . . . . . . . . . . . . . 279 7Independence of n random variables . . . . . . . . . . . . . . . 281 8Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281 9The one-dimensional normal distribution . . . . . . . . . . . . . 281 10 The multivariate normal distribution . . . . . . . . . . . . . . . 282 11 Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 285 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 286 13 The linear regression model287 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 287 2Affine minimum-trace unbiased estimation . . . . . . . . . . . . 288 3The Gauss-Markov theorem . . . . . . . . . . . . . . . . . . . . 289 4The method of least squares . . . . . . . . . . . . . . . . . . . . 292 5Aitken’s theorem . . . . . . . . . . . . . . . . . . . . . . . . . . 293 6Multicollinearity . . . . . . . . . . . . . . . . . . . . . . . . . . 295 7Estimable functions . . . . . . . . . . . . . . . . . . . . . . . . 297 8Linear constraints: the case M(R′) ⊂ M(X′) . . . . . . . . . . 299 9Linear constraints: the general case . . . . . . . . . . . . . . . . 302 10 Linear constraints: the case M(R′) ∩ M(X′) = {0} . . . . . . . 305 11 A singular variance matrix: the case M(X) ⊂ M(V ) . . . . . . 306 12 A singular variance matrix: the case r(X′V+X) = r(X) . . . . 308 13 A singular variance matrix: the general case, I . . . . . . . . . . 309 14 Explicit and implicit linear constraints . . . . . . . . . . . . . . 310 15 The general linear model, I . . . . . . . . . . . . . . . . . . . . 313 16 A singular variance matrix: the general case, II . . . . . . . . . 314 17 The general linear model, II . . . . . . . . . . . . . . . . . . . . 317 18 Generalized least squares . . . . . . . . . . . . . . . . . . . . . 318 19 Restricted least squares . . . . . . . . . . . . . . . . . . . . . . 319 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 321 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 322 14 Further topics in the linear model323 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323 2Best quadratic unbiased estimation of σ2. . . . . . . . . . . . 323 3The best quadratic and positive unbiased estimator of σ2. . . 324 4The best quadratic unbiased estimator of σ2. . . . . . . . . . . 326 5Best quadratic invariant estimation of σ2. . . . . . . . . . . . 329 6The best quadratic and positive invariant estimator of σ2. . . 330 7The best quadratic invariant estimator of σ2. . . . . . . . . . . 331 8Best quadratic unbiased estimation: multivariate normal case . 332 9Bounds for the bias of the least squares estimator of σ2, I . . . 335 10 Bounds for the bias of the least squares estimator of σ2, II . . . 336 11 The prediction of disturbances . . . . . . . . . . . . . . . . . . 338 12 Best linear unbiased predictors with scalar variance matrix . . 339 13 Best linear unbiased predictors with fixed variance matrix, I . . 341Contentsxi 14 Best linear unbiased predictors with fixed variance matrix, II . 344 15 Local sensitivity of the posterior mean . . . . . . . . . . . . . . 345 16 Local sensitivity of the posterior precision . . . . . . . . . . . . 347 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 348 Part Six — Applications to maximum likelihood estimation 15 Maximum likelihood estimation351 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351 2The method of maximum likelihood (ML) . . . . . . . . . . . . 351 3ML estimation of the multivariate normal distribution . . . . . 352 4Symmetry: implicit versus explicit treatment . . . . . . . . . . 354 5The treatment of positive definiteness . . . . . . . . . . . . . . 355 6The information matrix . . . . . . . . . . . . . . . . . . . . . . 356 7ML estimation of the multivariate normal distribution: distinct means . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357 8The multivariate linear regression model . . . . . . . . . . . . . 358 9The errors-in-variables model . . . . . . . . . . . . . . . . . . . 361 10 The non-linear regression model with normal errors . . . . . . . 364 11 Special case: functional independence of mean- and variance parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365 12 Generalization of Theorem 6 . . . . . . . . . . . . . . . . . . . 366 Miscellaneous exercises . . . . . . . . . . . . . . . . . . . . . . . . . . 368 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 370 16 Simultaneous equations371 1Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 371 2The simultaneous equations model . . . . . . . . . . . . . . . . 371 3The identification problem . . . . . . . . . . . . . . . . . . . . . 373 4Identification with linear constraints on B and Γ only . . . . . 375 5Identification with linear constraints on B,Γ and Σ . . . . . . . 375 6Non-linear constraints . . . . . . . . . . . . . . . . . . . . . . . 377 7Full-information maximum likelihood (FIML): the information matrix (general case) . . . . . . . . . . . . . . . . . . . . . . . . 378 8Full-information maximum likelihood (FIML): the asymptotic variance matrix (special case) . . . . . . . . . . . . . . . . . . . 380 9Limited-informationmaximumlikelihood(LIML): thefirst-order conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 383 10 Limited-information maximum likelihood (LIML): the informa- tion matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 386 11 Limited-information maximum likelihood (LIML): the asymp- totic variance matrix . . . . . . . . . . . . . . . . . . . . . . . . 388 Bibliographical notes . . . . . . . . . . . . . . . . . . . . . . . . . . . 393
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值