Maintaining G.S. Maddala’s brilliant expository style of cutting through the technical superstructure to reveal only essential details, while retaining the nerve centre of the subject matter, Professor Kajal Lahiri has brought forward this new edition of one of the most important textbooks in its field.  The new edition continues to provide a large number of worked examples, and some shorter data sets.  Further data sets and additional supplementary material to assist both the student and lecturer are available on the companion website www.wileyeurope.com/college/maddala
Les mer
Now in its fourth edition, this landmark text provides a fresh, accessible and well-written introduction to the subject. With a rigorous pedagogical framework, which sets it apart from comparable texts, the latest edition features an expanded website providing numerous real life data sets and examples.
Les mer
Foreword xvii Preface to the Fourth Edition xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics? 3 1.1 What is econometrics? 3 1.2 Economic and econometric models 4 1.3 The aims and methodology of econometrics 6 1.4 What constitutes a test of an economic theory? 8 CHAPTER 2 Statistical Background and Matrix Algebra 11 2.1 Introduction 11 2.2 Probability 12 2.3 Random variables and probability distributions 17 2.4 The normal probability distribution and related distributions 18 2.5 Classical statistical inference 21 2.6 Properties of estimators 22 2.7 Sampling distributions for samples from a normal population 26 2.8 Interval estimation 26 2.9 Testing of hypotheses 28 2.10 Relationship between confidence interval procedures and tests of hypotheses 31 2.11 Combining independent tests 32 CHAPTER 3 Simple Regression 59 3.1 Introduction 59 3.2 Specification of the relationships 61 3.3 The method of moments 65 3.4 The method of least squares 68 3.5 Statistical inference in the linear regression model 76 3.6 Analysis of variance for the simple regression model 83 3.7 Prediction with the simple regression model 85 3.8 Outliers 88 3.9 Alternative functional forms for regression equations 95 *3.10 Inverse prediction in the least squares regression model1 99 *3.11 Stochastic regressors 102 *3.12 The regression fallacy 102 CHAPTER 4 Multiple Regression 127 4.1 Introduction 127 4.2 A model with two explanatory variables 129 4.3 Statistical inference in the multiple regression model 134 4.4 Interpretation of the regression coefficients 143 4.5 Partial correlations and multiple correlation 146 4.6 Relationships among simple, partial, and multiple correlation coefficients 147 4.7 Prediction in the multiple regression model 153 4.8 Analysis of variance and tests of hypotheses 155 4.9 Omission of relevant variables and inclusion of irrelevant variables 160 4.10 Degrees of freedom and R2 165 4.11 Tests for stability 169 4.12 The LR, W, and LM tests 176 Part II Violation of the Assumptions of the Basic Regression Model 209 CHAPTER 5 Heteroskedasticity 211 5.1 Introduction 211 5.2 Detection of heteroskedasticity 214 5.3 Consequences of heteroskedasticity 219 5.4 Solutions to the heteroskedasticity problem 221 5.5 Heteroskedasticity and the use of deflators 224 5.6 Testing the linear versus log-linear functional form 228 CHAPTER 6 Autocorrelation 239 6.1 Introduction 239 6.2 The Durbin–Watson test 240 6.3 Estimation in levels versus first differences 242 6.4 Estimation procedures with autocorrelated errors 246 6.5 Effect of AR(1) errors on OLS estimates 250 6.6 Some further comments on the DW test 254 6.7 Tests for serial correlation in models with lagged dependent variables 257 6.8 A general test for higher-order serial correlation: The LM test 259 6.9 Strategies when the DW test statistic is significant 261 *6.10 Trends and random walks 266 *6.11 ARCH models and serial correlation 271 6.12 Some comments on the DW test and Durbin’s h-test and t-test 272 CHAPTER 7 Multicollinearity 279 7.1 Introduction 279 7.2 Some illustrative examples 280 7.3 Some measures of multicollinearity 283 7.4 Problems with measuring multicollinearity 286 7.5 Solutions to the multicollinearity problem: Ridge regression 290 7.6 Principal component regression 292 7.7 Dropping variables 297 7.8 Miscellaneous other solutions 300 CHAPTER 8 Dummy Variables and Truncated Variables 313 8.1 Introduction 313 8.2 Dummy variables for changes in the intercept term 314 8.3 Dummy variables for changes in slope coefficients 319 8.4 Dummy variables for cross-equation constraints 322 8.5 Dummy variables for testing stability of regression coefficients 324 8.6 Dummy variables under heteroskedasticity and autocorrelation 327 8.7 Dummy dependent variables 329 8.8 The linear probability model and the linear discriminant function 329 8.9 The probit and logit models 333 8.10 Truncated variables: The tobit model 343 CHAPTER 9 Simultaneous Equation Models 355 9.1 Introduction 355 9.2 Endogenous and exogenous variables 357 9.3 The identification problem: Identification through reduced form 357 9.4 Necessary and sufficient conditions for identification 362 9.5 Methods of estimation: The instrumental variable method 365 9.6 Methods of estimation: The two-stage least squares method 371 9.7 The question of normalization 378 *9.8 The limited-information maximum likelihood method 379 *9.9 On the use of OLS in the estimation of simultaneous equation models 380 *9.10 Exogeneity and causality 386 9.11 Some problems with instrumental variable methods 392 CHAPTER 10 Diagnostic Checking, Model Selection, and Specification Testing 401 10.1 Introduction 401 10.2 Diagnostic tests based on least squares residuals 402 10.3 Problems with least squares residuals 404 10.4 Some other types of residual 405 10.5 DFFITS and bounded influence estimation 411 10.6 Model selection 414 10.7 Selection of regressors 419 10.8 Implied F-ratios for the various criteria 423 10.9 Cross-validation 427 10.10 Hausman’s specification error test 428 10.11 The Plosser–Schwert–White differencing test 435 10.12 Tests for nonnested hypotheses 436 10.13 Nonnormality of errors 440 10.14 Data transformations 441 CHAPTER 11 Errors in Variables 451 11.1 Introduction 451 11.2 The classical solution for a single-equation model with one explanatory variable 452 11.3 The single-equation model with two explanatory variables 455 11.4 Reverse regression 463 11.5 Instrumental variable methods 465 11.6 Proxy variables 468 11.7 Some other problems 471 Part III Special Topics 479 CHAPTER 12 Introduction to Time-Series Analysis 481 12.1 Introduction 481 12.2 Two methods of time-series analysis: Frequency domain and time domain 482 12.3 Stationary and nonstationary time series 482 12.4 Some useful models for time series 485 12.5 Estimation of AR, MA, and ARMA models 492 12.6 The Box–Jenkins approach 496 12.7 R2 measures in time-series models 503 CHAPTER 13 Models of Expectations and Distributed Lags 509 13.1 Models of expectations 509 13.2 Naive models of expectations 510 13.3 The adaptive expectations model 512 13.4 Estimation with the adaptive expectations model 514 13.5 Two illustrative examples 516 13.6 Expectational variables and adjustment lags 520 13.7 Partial adjustment with adaptive expectations 524 13.8 Alternative distributed lag models: Polynomial lags 526 13.9 Rational lags 533 13.10 Rational expectations 534 13.11 Tests for rationality 536 13.12 Estimation of a demand and supply model under rational expectations 538 13.13 The serial correlation problem in rational expectations models 544 CHAPTER 14 Vector Autoregressions, Unit Roots, and Cointegration 551 14.1 Introduction 551 14.2 Vector autoregressions 551 14.3 Problems with VAR models in practice 553 14.4 Unit roots 554 14.5 Unit root tests 555 14.6 Cointegration 563 14.7 The cointegrating regression 564 14.8 Vector autoregressions and cointegration 567 14.9 Cointegration and error correction models 571 14.10 Tests for cointegration 571 14.11 Cointegration and testing of the REH and MEH 572 14.12 A summary assessment of cointegration 574 CHAPTER 15 Panel Data Analysis 583 15.1 Introduction 583 15.2 The LSDV or fixed effects model 584 15.3 The random effects model 586 15.4 Fixed effects versus random effects 589 15.5 Dynamic panel data models 591 15.6 Panel data models with correlated effects and simultaneity 593 15.7 Errors in variables in panel data 595 15.8 The SUR model 597 15.9 The random coefficient model 597 CHAPTER 16 Small-Sample Inference: Resampling Methods 601 16.1 Introduction 601 16.2 Monte Carlo methods 602 16.3 Resampling methods: Jackknife and bootstrap 603 16.4 Bootstrap confidence intervals 605 16.5 Hypothesis testing with the bootstrap 606 16.6 Bootstrapping residuals versus bootstrapping the data 607 16.7 Non-IID errors and nonstationary models 607 Appendix 611 Index 621
Les mer
Maintaining G.S. Maddala’s brilliant expository style of cutting through the technical superstructure to reveal only essential details, while retaining the nerve centre of the subject matter, Professor Kajal Lahiri has brought forward this new edition of one of the most important textbooks in its field. The new edition continues to provide a large number of worked examples, and some shorter data sets.  Further data sets and additional supplementary material to assist both the student and lecturer are available on the companion website www.wileyeurope.com/college/maddala New features for the fourth edition: Chapters 5 and 6, on Heteroscedasticity and Autocorrelation, now reflect the latest professional practice in dealing with these common variations of the basic regression model.Chapter 10 includes extensive discussion on diagnostic checking in linear models, various nested and non-nested model selection procedures, specification testing, data transformations, and tests for non-normality.The first three chapters of Part III cover an introduction to time-series analysis, including the Box–Jenkins approach, forecasting and seasonality, models of expectations and distributed lag models, and vector auto-regressions, unit roots, and cointegration.Chapters 15 and 16 cover, respectively, the latest developments in panel data analysis and various re-sampling methods for use in small sample inference.
Les mer
Part I: Introduction and the Linear Regression Model Chapter 1             What is Econometrics? 1.1          What is Econometrics? 1.2          Economic and Econometric Models 1.3          The Aims and Methodology of Econometrics 1.4          What Constitutes a Test of an Economic Theory? Summary and an Outline of the Book References Chapter 2              Statistical Background and Matrix Algebra 2.1          Introduction 2.2          Probability Addition Rules of Probability Conditional Probability and the Multiplication Rule Bayes? Theorem Summation and Product Operations 2.3          Random Variables and Probability Distributions Joint, Marginal, and Conditional Distributions Illustrative Example 2.4          The Normal Probability Distribution and Related Distributions The Normal Distribution Related Distributions 2.5          Classical Statistical Inference Point Estimation 2.6          Properties of Estimators Unbiasedness Efficiency Consistency Other Asymptotic Properties 2.7          Sampling Distributions for Samples from a Normal Population 2.8          Interval Estimation 2.9          Testing of Hypotheses 2.10        Relationship between Confidence Interval Procedures and Tests of Hypotheses 2.11        Combining Independent Tests Summary Exercises Appendix: Matrix Algebra Exercises on Matrix Algebra References Chapter 3             Simple Regression 3.1          Introduction Example 1: Simple Regression Example 2: Multiple Regression 3.2          Specification of the Relationships 3.3          The Method of Moments Illustrative Example 3.4          The Method of Least Squares Reverse Regression Illustrative Example 3.5          Statistical Inference in the Linear Regression Model Illustrative Example Confidence Intervals for a, b, and s2 Testing of Hypotheses Example of Comparing Test Scores from the GRE and GMAT Tests Regression with No Constant Term 3.6          Analysis of Variance for the Simple Regression Model 3.7          Prediction with the Simple Regression Model Prediction of Expected Values Illustrative Example 3.8          Outliers Some Illustrative Examples 3.9          Alternative Functional Forms for Regression Equations Illustrative Example 3.10        Inverse Prediction in the Least Squares Regression Model 3.11        Stochastic Regressors 3.12        The Regression Fallacy The Bivariate Normal Distribution Galton?s Result and the Regression Fallacy A Note on the Term ?Regression? Summary Exercises Appendix: Proofs References Chapter 4             Multiple Regression 4.1          Introduction 4.2          A Model with Two Explanatory Variables The Least Squares Method Illustrative Example 4.3          Statistical Inference in the Multiple Regression Model Illustrative Example Formulas for the General Case of k Explanatory Variables Some Illustrative Examples 4.4          Interpretation of the Regression Coefficients Illustrative Example 4.5          Partial Correlations and Multiple Correlation 4.6          Relationships among Simple, Partial, and Multiple Correlation Coefficients Two Illustrative Examples 4.7          Prediction in the Multiple Regression Model Illustrative Example 4.8          Analysis of Variance and Tests of Hypotheses Nested and Nonnested Hypotheses Tests for Linear Functions of Parameters Illustrative Example 4.9          Omission of Relevant Variables and Inclusion of Irrelevant Variables Omission of Relevant Variables Example 1: Demand for Food in the United States Example 2: Production Functions and Management Bias Inclusion of Irrelevant Variables 4.10        Degrees of Freedom and 4.11        Tests for Stability The Analysis of Variance Test Example 1: Stability of the Demand for Food Function Example 2: Stability of Production Functions Predictive Tests for Stability Illustrative Example 4.12        The LR, W, and LM Tests Illustrative Example Summary Exercises Appendix 4.1: The Multiple Regression Model in Matrix Notation Appendix 4.2: Nonlinear Regressions Appendix 4.3: Large-Sample Theory Data Sets References Part II: Violation of the Assumptions of the Basic Model Chapter 5             Heteroskedasticity 5.1          Introduction Illustrative Example 5.2          Detection of Heteroskedasticity Illustrative Example Some Other Tests Illustrative Example An Intuitive Justification for the Breusch?Pagan Test 5.3          Consequences of Heteroskedasticity Estimation of the Variance of the OLS Estimator under Heteroskedasticity 5.4          Solutions to the Heteroskedasticity Problem Illustrative Example 5.5          Heteroskedasticity and the Use of Deflators Illustrative Example: The Density Gradient Model 5.6          Testing the Linear versus Log-Linear Functional Form The Box?Cox Test The BM Test The PE Test Summary Exercises Appendix: Generalized Least Squares References Chapter 6             Autocorrelation 6.1          Introduction 6.2          The Durbin?Watson Test Illustrative Example 6.3          Estimation in Levels versus First Differences Some Illustrative Examples 6.4          Estimation Procedures with Autocorrelated Errors Iterative Procedures Grid-Search Procedures 6.5          Effect of AR(1) Errors on OLS Estimates 6.6          Some Further Comments on the DW Test The von Neumann Ratio The Berenblut?Webb Test 6.7          Tests for Serial Correlation in Models with Lagged Dependent Variables Durbin?s h-Test Durbin?s Alternative Test Illustrative Example 6.8          A General Test for Higher-Order Serial Correlation: The LM Test 6.9          Strategies When the DW Test Statistic is Significant Errors Not AR(1) Autocorrelation Caused by Omitted Variables Serial Correlation Due to Misspecified Dynamics The Wald Test Illustrative Example 6.10        Trends and Random Walks Spurious Trends Differencing and Long-Run Effects: The Concept of Cointegration 6.11        ARCH Models and Serial Correlation 6.12        Some Comments on the DW Test and Durbin?s h-Test and t-Test Summary Exercises References Chapter 7             Multicollinearity 7.1          Introduction 7.2          Some Illustrative Examples 7.3          Some Measures of Multicollinearity 7.4          Problems with Measuring Multicollinearity 7.5          Solutions to the Multicollinearity Problem: Ridge Regression 7.6          Principal Component Regression 7.7          Dropping Variables 7.8          Miscellaneous Other Solutions Using Ratios or First Differences Using Extraneous Estimates Getting More Data Summary Exercises Appendix: Linearly Dependent Explanatory Variables References                 Chapter 8             Dummy Variables and Truncated Variables 8.1          Introduction 8.2          Dummy Variables for Changes in the Intercept Term Illustrative Example Two More Illustrative Examples 8.3          Dummy Variables for Changes in Slope Coefficients 8.4          Dummy Variables for Cross-Equation Constraints 8.5          Dummy Variables for Testing Stability of Regression Coefficients 8.6          Dummy Variables under Heteroskedasticity and Autocorrelation 8.7          Dummy Dependent Variables 8.8          The Linear Probability Model and the Linear Discriminant Function The Linear Probability Model The Linear Discriminant Function 8.9          The Probit and Logit Models Illustrative Example The Problem of Disproportionate Sampling Prediction of Effects of Changes in the Explanatory Variables Measuring Goodness of Fit 8.10        Illustrative Example 8.11        Truncated Variables: The Tobit Model Some Examples Method of Estimation Limitations of the Tobit Model The Truncated Regression Model Summary Exercises References Chapter 9             Simultaneous Equations Models 9.1          Introduction 9.2          Endogenous and Exogenous Variables 9.3          The Identification Problem: Identification through Reduced Form Illustrative Example 9.4          Necessary and Sufficient Conditions for Identification Illustrative Example 9.5          Methods of Estimation: The Instrumental Variable Method Measuring R2 Illustrative Example 9.6          Methods of Estimation: The Two-Stage Least Squares Method Computing Standard Errors Illustrative Example 9.7          The Question of Normalization 9.8          The Limited-Information Maximum Likelihood Method Illustrative Example 9.9          On the Use of OLS in the Estimation of Simultaneous Equations Models Working?s Concept of Identification Recursive Systems Estimation of Cobb?Douglas Production Functions 9.10        Exogeneity and Causality Weak Exogeneity Superexogeneity Strong Exogeneity Granger Causality Granger Causality and Exogeneity Tests for Exogeneity 9.11        Some Problems with Instrumental Variable Methods Summary Exercises Appendix References Chapter 10           Diagnostic Checking, Model Selection, and Specification Testing 10.1        Introduction 10.2        Diagnostic Tests Based on Least Squares Residuals Tests for Omitted Variables Tests for ARCH Effects 10.3        Problems with Least Squares Residuals 10.4        Some Other Types of Residuals Predicted Residuals and Studentized Residuals Dummy Variable Method for Studentized Residuals BLUS Residuals Recursive Residuals Illustrative Example 10.5        DFFITS and Bounded Influence Estimation Illustrative Example 10.6        Model Selection Hypothesis-Testing Search Interpretive Search Simplification Search Proxy Variable Search Data Selection Search Post-Data Model Construction Hendry?s Approach to Model Selection 10.7        Selection of Regressors Theil?s   Criterion Criteria Based on Minimizing the Mean-Squared Error of Prediction Akaike?s Information Criterion 10.8        Implied F-Ratios for the Various Criteria Bayes? Theorem and Posterior Odds for Model Selection 10.9        Cross-Validation 10.10      Hausman?s Specification Error Test An Application: Testing for Errors in Variables or Exogeneity Some Illustrative Examples An Omitted Variable Interpretation of the Hausman Test 10.11      The Plosser?Schwert?White Differencing Test 10.12      Tests for Nonnested Hypotheses The Davidson and MacKinnon Test The Encompassing Test A Basic Problem in Testing Nonnested Hypotheses Hypothesis Testing versus Model Selection as a Research Strategy 10.13      Nonnormality of Errors Tests for Normality 10.14      Data Transformations Summary Exercises Appendix References Chapter 11           Errors in Variables 11.1        Introduction 11.2        The Classical Solution for a Single-Equation Model with One Explanatory Variable 11.3        The Single-Equation Model with Two Explanatory Variables Two Explanatory Variables: One Measured with Error Illustrative Example Two Explanatory Variables: Both Measured with Error 11.4        Reverse Regression 11.5        Instrumental Variable Methods 11.6        Proxy Variables Coefficient of the Proxy Variable 11.7        Some Other Problems The Case of Multiple Equations Correlated Errors Summary Exercises References Part III: Special Topics Chapter 12           Introduction to Time-Series Analysis 12.1        Introduction 12.2        Two Methods of Time-Series Analysis: Frequency Domain and Time Domain 12.3        Stationary and Nonstationary Time Series Strict Stationarity Weak Stationarity Properties of Autocorrelation Function Nonstationarity 12.4        Some Useful Models for Time Series Purely Random Process Random Walk Moving Average Process Autoregressive Process Autoregressive Moving Average Process Autoregressive Integrated Moving Average Process 12.5        Estimation of AR, MA, and ARMA Models Estimation of MA Models Estimation of ARMA Models Residuals from the ARMA Models Testing Goodness of Fit 12.6        The Box?Jenkins Approach Forecasting from Box?Jenkins Models Illustrative Example Trend Elimination: The Traditional Method A Summary Assessment Seasonality in the Box?Jenkins Modeling 12.7        R 2 Measures in Time-Series Models Summary Exercises Data Sets References Chapter 13           Models of Expectations and Distributed Lags 13.1        Models of Expectations 13.2        Naive Models of Expectations 13.3        The Adaptive Expectations Model 13.4        Estimation with the Adaptive Expectations Model Estimation in the Autoregressive Form Estimation in the Distributed Lag Form 13.5        Two Illustrative Examples 13.6        Expectational Variables and Adjustment Lags 13.7        Partial Adjustment with Adaptive Expectations 13.8        Alternative Distributed Lag Models: Polynomial Lags Finite Lags: The Polynomial Lag Illustrative Example Choosing the Degree of the Polynomial 13.9        Rational Lags 13.10      Rational Expectations 13.11      Tests for Rationality 13.12      Estimation of a Demand and Supply Model Under Rational Expectations Case 1 Case 2 Illustrative Example 13.13      The Serial Correlation Problem in Rational Expectations Models Summary Exercises References Chapter 14           Vector Autoregressions, Unit Roots, and Cointegration 14.1        Introduction 14.2        Vector Autoregressions 14.3        Problems with VAR Models in Practice 14.4        Unit Roots 14.5        Unit Root Tests The Dickey?Fuller Tests The Serial Correlation Problem The Low Power of Unit Root Tests The DF-GLS Test What are the Null and Alternative Hypotheses in Unit Root Tests? Tests with Stationarity as Null Confirmatory Analysis Panel Data Unit Root Tests Structural Change and Unit Roots 14.6        Cointegration 14.7        The Cointegrating Regression 14.8        Vector Autoregressions and Cointegration 14.9        Cointegration and Error Correction Models 14.10      Tests for Cointegration 14.11      Cointegration and Testing of the REH and MEH 14.12      A Summary Assessment of Cointegration Summary Exercises References Chapter 15           Panel Data Analysis 15.1        Introduction 15.2        The LSDV or Fixed Effects Model Illustrative Example: Fixed Effect Estimation 15.3        The Random Effects Model 15.4        Fixed Effects versus Random Effects Hausman Test Breusch and Pagan Test Tests for Serial Correlation 15.5        Dynamic Panel Data Models 15.6        Panel Data Models with Correlated Effects and Simultaneity 15.7        Errors in Variables in Panel Data 15.8        The SUR Model 15.9        The Random Coefficient Model Summary References Chapter 16           Small-Sample Inference: Resampling Methods 16.1        Introduction 16.2        Monte Carlo Methods More Efficient Monte Carlo Methods Response Surfaces 16.3        Resampling Methods: Jackknife and Bootstrap Some Illustrative Examples Other Issues Relating to the Bootstrap 16.4        Bootstrap Confidence Intervals 16.5        Hypothesis Testing with the Bootstrap 16.6        Bootstrapping Residuals versus Bootstrapping the Data 16.7        Non-IID Errors and Nonstationary Models Heteroskedasticity and Autocorrelation Unit Root Tests Based on the Bootstrap Cointegration Tests 16.8        Miscellaneous Other Applications Summary References Appendix Index
Les mer

Produktdetaljer

ISBN
9780470015124
Publisert
2009-10-16
Utgave
4. utgave
Utgiver
Vendor
John Wiley & Sons Inc
Vekt
1191 gr
Høyde
238 mm
Bredde
190 mm
Dybde
38 mm
Aldersnivå
P, 06
Språk
Product language
Engelsk
Format
Product format
Heftet
Antall sider
656

Biographical note

G.S.Maddala was one of the leading figures in field of econometrics for more than 30 years until he passed away in 1999. At the time of his death, he held the University Eminent Scholar Professorship in the Department of Economics at Ohio State University. His previous affiliations include Stanford University, University of Rochester and University of Florida.

Kajal Lahiri is Distinguished Professor of Economics, and Health Policy, and Management and Behaviour at the State University of New York, Albany where he is also Director of the Econometric Research Institute. Professor Lahiri is an Honorary Fellow of the International Institute of Forecasters.