In several chapters we have discussed goodness-of-fit tests to assess the

Similar documents
AIC deals with the trade-off between the complexity of the model and the goodness of fit of the model.

MONETARY POLICY AND HOUSING MARKET: COINTEGRATION APPROACH

AIC and BIC are each methods of comparing models fit via maximum likelihood. In each case, a larger value is better.

CABARRUS COUNTY 2016 APPRAISAL MANUAL

IREDELL COUNTY 2015 APPRAISAL MANUAL

Initial sales ratio to determine the current overall level of value. Number of sales vacant and improved, by neighborhood.

EXPLANATION OF MARKET MODELING IN THE CURRENT KANSAS CAMA SYSTEM

A. K. Alexandridis University of Kent. D. Karlis Athens University of Economics and Business. D. Papastamos Eurobank Property Services S.A.

The Improved Net Rate Analysis

AVM Validation. Evaluating AVM performance

Hedonic Pricing Model Open Space and Residential Property Values

Volume 35, Issue 1. Real Interest Rate and House Prices in Malaysia: An Empirical Study

BUSI 398 Residential Property Guided Case Study

Using Hedonics to Create Land and Structure Price Indexes for the Ottawa Condominium Market

86M 4.2% Executive Summary. Valuation Whitepaper. The purposes of this paper are threefold: At a Glance. Median absolute prediction error (MdAPE)

Sponsored by a Grant TÁMOP /2/A/KMR Course Material Developed by Department of Economics, Faculty of Social Sciences, Eötvös Loránd

Following is an example of an income and expense benchmark worksheet:

Mass appraisal Educational offerings and Designation Requirements. designations provide a portable measurement of your capabilities

Volume Title: Well Worth Saving: How the New Deal Safeguarded Home Ownership

Gregory W. Huffman. Working Paper No. 01-W22. September 2001 DEPARTMENT OF ECONOMICS VANDERBILT UNIVERSITY NASHVILLE, TN 37235

Definitions ad valorem tax Adaptive Estimation Procedure (AEP) - additive model - adjustments - algorithm - amenities appraisal appraisal schedules

The Housing Price Bubble, Monetary Policy, and the Foreclosure Crisis in the U.S.

Property Tax in Upstate New York

Metro Boston Perfect Fit Parking Initiative

Comparables Sales Price (Old Version)

Course Commerical/Industrial Modeling Concepts Learning Objectives

Over the past several years, home value estimates have been an issue of

The Analytic Hierarchy Process. M. En C. Eduardo Bustos Farías

Basic Appraisal Procedures

The Effect of Relative Size on Housing Values in Durham

DATA APPENDIX. 1. Census Variables

Washington Department of Revenue Property Tax Division. Valid Sales Study Kitsap County 2015 Sales for 2016 Ratio Year.

Valuing Land in Dispute Resolution: Using Coefficient of Variation to Determine Unit of Measurement

Intangibles CHAPTER CHAPTER OBJECTIVES. After careful study of this chapter, you will be able to:

Ad-valorem and Royalty Licensing under Decreasing Returns to Scale

Trends in Affordable Home Ownership in Calgary

A Real-Option Based Dynamic Model to Simulate Real Estate Developer Behavior

What Factors Determine the Volume of Home Sales in Texas?

Hunting the Elusive Within-person and Between-person Effects in Random Coefficients Growth Models

The purpose of the appraisal was to determine the value of this six that is located in the Town of St. Mary s.

Can the coinsurance effect explain the diversification discount?

Comparison of Dynamics in the Korean Housing Market Based on the FDW Model for the Periods Before and After the Macroeconomic Fluctuations

The Impact of the New Revenue Standard on Real Estate Sales

On the Disutility and Discounting of Imprisonment and the Theory of Deterrence

On the Relationship between Track Geometry Defects and Development of Internal Rail Defects

Measuring Urban Commercial Land Value Impacts of Access Management Techniques

Metro Vancouver Purpose-Built Rental Housing Inventory and Risk Analysis. Profile for the District of West Vancouver

GASB 69: Government Combinations

Explanation of the Analysis Format

Use of Comparables. Claims Prevention Bulletin [CP-17-E] March 1996

Demonstration Properties for the TAUREAN Residential Valuation System

DIRECTIVE # This Directive Supersedes Directive # and #92-003

absorption rate ad valorem appraisal broker price opinion capital gain

Economic Impacts of MLS Home Sales and Purchases in Canada and the Provinces

SAS at Los Angeles County Assessor s Office

Tangible Personal Property Summation Valuation Procedures

RE: Proposed Accounting Standards Update, Leases (Topic 842): Targeted Improvements (File Reference No )

Proving Depreciation

COMPARISON OF THE LONG-TERM COST OF SHELTER ALLOWANCES AND NON-PROFIT HOUSING

On the Choice of Tax Base to Reduce. Greenhouse Gas Emissions in the Context of Electricity. Generation

Section 12 Accounting for Leases Accounting by the Lessor and Lessee

Part 2. Measures of Central Tendency: Mean, Median, and Mode

Hennepin County Economic Analysis Executive Summary

Online Appendix "The Housing Market(s) of San Diego"

The Proposal of Cadastral Value Determination Based on Artificial Intelligence

p URCHASE of development rights

ARIZONA TAX COURT TX /18/2006 HONORABLE MARK W. ARMSTRONG

Supplemental Handout

UNIVALENCE CRITERIA OF CERTAIN INTEGRAL OPERATOR (Kriterium Univalen bagi Pengoperasi Kamiran Tertentu)

Estimating User Accessibility Benefits with a Housing Sales Hedonic Model

Applying IFRS. Impairment considerations for the new leasing standard. November 2018

LITIGATING IN A MASS APPRAISAL ENVIRONMENT

increases. See 7.09 supra discussing the issues inherent with the sum of the demised and demisable premises in a building.

Estimating the Value of Foregone Rights on Land. A Working Paper Prepared for the Vermillion River Watershed Joint Powers Organization 1.

Course Number Course Title Course Description

Litigation Valuation REPORT. Rules of thumb: Do they have a place in business valuation? When a business is being valued, consider the real estate

Solutions to Questions

City of Puyallup. Parks Impact Fee Study

Village of Scarsdale

Objectives of Housing Task Force: Some Background

International Valuation Standards Board 15 Feb Moorgate London EC2R 6PP United Kingdom

Relationship of age and market value of office buildings in Tirana City

Department of Economics Working Paper Series

ONLINE APPENDIX "Foreclosures, House Prices, and the Real Economy" Atif Mian Amir Sufi Francesco Trebbi [NOT FOR PUBLICATION]

American Community Survey 5-Year Estimates

The capitalization rate is essential to any analysis through the income

BUSI 499 Income Property Guided Case Study

Volume Title: Accelerated Depreciation in the United States, Volume URL:

Tobin s q what to do?

Chapter 13. The Market Approach to Value

Course Residential Modeling Concepts

An analysis of affinity programs: the case of real estate brokerage participation

The Long-Run Relationship between House Prices and Inflation in South Africa: An ARDL Approach *

RE: Request for Comments on the Exposure Draft The Valuation of Forests dated November 16, 2012

Sales Ratio: Alternative Calculation Methods

How to Read a Real Estate Appraisal Report

American Community Survey 5-Year Estimates

3rd Meeting of the Housing Task Force

Journal of Business & Economics Research Volume 1, Number 9

A Quantitative Approach to Gentrification: Determinants of Gentrification in U.S. Cities,

Transcription:

The Basics of Financial Econometrics: Tools, Concepts, and Asset Management Applications. Frank J. Fabozzi, Sergio M. Focardi, Svetlozar T. Rachev and Bala G. Arshanapalli. 2014 John Wiley & Sons, Inc. Published 2014 by John Wiley & Sons, Inc. APPENDIX E Model Selection Criterion: AIC and BIC In several chapters we have discussed goodness-of-fit tests to assess the performance of a model with respect to how well it explains the data. However, suppose we want to select from among several candidate models. What criterion can be used to select the best model? In choosing a criterion for model selection, one accepts the fact that models only approximate reality. Given a set of data, the objective is to determine which of the candidate models best approximates the data. This involves trying to minimize the loss of information. Because the field of information theory is used to quantify or measure the expected value of information, the information-theoretic approach is used to derive the two most commonly used criteria in model selection the Akaike information criterion and the Bayesian information criterion. 1 These two criteria, as described in this appendix, can be used for the selection of econometric models. 2 1 There are other approaches that have been developed. One approach is based on the theory of learning, the Vapnik-Chervonenkis (VC) theory of learning. This approach offers a complex theoretical framework for learning that, when applicable, is able to give precise theoretical bounds to the learning abilities of models. Though its theoretical foundation is solid, the practical applicability of the VC theory is complex. It has not yet found a broad following in financial econometrics. See Vladimir N. Vapnik, Statistical Learning Theory (New York: John Wiley & Sons, 1998). 2 For a further discussion of these applications of AIC and BIC, see Herman J. Bierens, Information Criteria and Model Selection, Pennsylvania State University, March 12, 2006, working paper. In addition to the AIC and BIC, Bierens discusses another criterion, the Hannan-Quinn criterion, in E. J. Hannan and B. G. Quinn, The Determination of the Order of an Autoregression, Journal of the Royal Statistical Society B, no. 41 (1979): 190 195. 399

400 THE BASICS OF FINANCIAL ECONOMETRICS AKAIKE INFORMATION CRITERION In 1951, Kullback and Leibler developed a measure to capture the information that is lost when approximating reality; that is, the Kullback and Leibler measure is a criterion for a good model that minimizes the loss of information. 3 Two decades later, Akaike established a relationship between the Kullback-Leibler measure and maximum likelihood estimation method an estimation method used in many statistical analyses as described in Chapter 13 to derive a criterion (i.e., formula) for model selection. 4 This criterion, referred to as the Akaike information criterion (AIC), is generally considered the first model selection criterion that should be used in practice. The AIC is AIC = 2log L(ˆ) θ + 2k where θ = the set (vector) of model parameters L(ˆ) θ = the likelihood of the candidate model given the data when evaluated at the maximum likelihood estimate of θ k = the number of estimated parameters in the candidate model The AIC in isolation is meaningless. Rather, this value is calculated for every candidate model and the best model is the candidate model with the smallest AIC. Let s look at the two components of the AIC. The first component, 2log L(ˆ), θ is the value of the likelihood function, log L(ˆ), θ which is the probability of obtaining the data given the candidate model. Since the likelihood function s value is multiplied by 2, ignoring the second component, the model with the minimum AIC is the one with the highest value for the likelihood function. However, to this first component we add an adjustment based on the number of estimated parameters. The more parameters, the greater the amount added to the first component, increasing the value for the AIC and penalizing the model. Hence, there is a trade-off: the better fit, created by making a model more complex by requiring more parameters, must be considered in light of the penalty imposed by adding more parameters. This is why the second component of the AIC is thought of in terms of a penalty. 3 S. Kullback and R. A. Leibler, On Information and Sufficiency, Annals of Mathematical Statistics 22, no. 1 (1951): 79 86. 4 Hirotugu Akaike, Information Theory and an Extension of the Maximum Likelihood Principle, in Second International Symposium on Information Theory, ed. B. N. Petrov and F. Csake (Budapest: Akademiai Kiado, 1973), 267 281; and Hirotugu Akaike, A New Look at the Statistical Model Identification, I.E.E.E. Transactions on Automatic Control, AC 19, (1974): 716 723.

Model Selection Criterion: AIC and BIC 401 For small sample sizes, the second-order Akaike information criterion (AIC c ) should be used in lieu of the AIC described earlier. The AIC c is AICc = 2log L(ˆ) θ + 2 k+ (2k+ 1) / ( n k 1) where n is the number of observations. 5 A small sample size is when n/k is less than 40. Notice as the n increases, the third term in AIC c approaches zero and will therefore give the same result as AIC. AIC c has also been suggested to be used instead of AIC when n is small or k is large. 6 It has been suggested, for example, that in selecting the orders of an ARMA, as we described in Chapter 9, the AIC c be used. 7 Typically, to assess the strength of evidence for the each candidate model, two measures can be used: 1. The delta AIC 2. The Akaike weights Consider first the delta AIC measure assuming there are M candidate models. An AIC can be calculated for each candidate model, denoted by AIC m (m = 1,..., M). The AIC with the minimum value, denoted by AIC *, is then the best model. The delta AIC for the mth candidate model, denoted by Δ m, is simply the difference between the AIC m and AIC *. This difference is then used as follows to determine the level of support for each candidate model. If the delta AIC is Less than 2, this indicates there is substantial evidence to support the candidate model (i.e., the candidate model is almost as good as the best model). Between 4 and 7, this indicates that the candidate model has considerably less support. Greater than 10, this indicates that there is essentially no support for the candidate model (i.e., it is unlikely to be the best model). 8 The above values for the computed delta AICs are merely general rules of thumb. Because the magnitude of the delta AIC is not meaningful in itself, to measure the strength of evidence for a candidate model we are interested 5 Clifford M. Hurvich and Chih-Ling Tsai, Regression and Time Series Model Selection in Small Samples, Biometrika 76, no. 2 ( June 1989): 297 307. 6 Kenneth P. Burnham and David R. Anderson, Model Selection and Multimodel Inference: A Practical Information-Theoretic Approach, 2nd ed. (New York: Springer- Verlag, 2002). 7 Peter J. Brockwell and Richard A. Davis, Time Series: Theory and Methods, 2nd ed. (New York: Springer-Verlag, 2009), 273. 8 Burnham and Anderson, Model Selection and Multimodel Inference, 70.

402 THE BASICS OF FINANCIAL ECONOMETRICS in the relative value of the delta AIC. The Akaike weights, denoted by w m, are obtained by normalizing the relative likelihood values. That is, they are the ratios of a candidate model s delta AIC relative to the sum of the delta AICs for all candidate models, shown as follows: exp( 0.5 Δm) wm = M exp( 0.5 Δ ) j= 1 The interpretation of this measure of strength of each candidate model given the data is the following: the Akaike weights are the probability that the candidate model is the best among the set of candidate models. For example, if a candidate model has an Akaike weight of 0.60, this means that given the data, the candidate model has a 60% probability of being the best one. Further information can be obtained by calculating the ratio of Akaike weights for different candidate models to determine to what extent one candidate model is better than another candidate model. These measures, called evidence ratios, can be used to compare, for example, the best model versus a candidate model. For example, if the evidence ratio computed as the ratio of the best model to some candidate model is 1.8, then this can be interpreted as the best model being 1.8 times more likely than that candidate model of being the best model. What is the difference between the AIC and hypothesis tests in model selection described in Chapters 3 and 4 where we described statistical tests for various regression models and the use of stepwise regressions? The difference is that in those earlier chapters, the tests used for model selection are hypothesis tests where at a certain level of confidence an independent variable would be included or excluded from the model. In contrast, model selection applying AIC is based on the strength of the evidence and provides for each of the candidate models a measure of uncertainty. What is important to emphasize is that the AIC might identify which model is best among the candidate models but that does not mean that any of the candidate models do a good job of explaining the data. BAYESIAN INFORMATION CRITERION The Bayesian information criterion (BIC), proposed by Schwarz 9 and hence also referred to as the Schwarz information criterion and Schwarz Bayesian j 9 Gideon Schwarz, Estimating the Dimension of a Model, Annals of Statistics 6 (1978): 461 464. The purpose of the BIC is to provide an asymptotic approximation to a transformation of the candidate model s Bayesian posterior probability.

Model Selection Criterion: AIC and BIC 403 information criterion, is another model selection criterion based on information theory but set within a Bayesian context. The difference between the BIC and the AIC is the greater penalty imposed for the number of parameters by the former than the latter. Burnham and Anderson provide theoretical arguments in favor of the AIC, particularly the AIC c over the BIC. 10 Moreover, in the case of multivariate regression analysis, Yang explains why AIC is better than BIC in model selection. 11 The BIC is computed as follows: BIC = 2log L(ˆ) θ + klogn where the terms above are the same as described in our description of the AIC. The best model is the one that provides the minimum BIC, denoted by BIC*. Like delta AIC for each candidate model, we can compute delta BIC = BIC m BIC*. Given M models, the magnitude of the delta BIC can be interpreted as evidence against a candidate model being the best model. The rules of thumb are 12 Less than 2, it is not worth more than a bare mention. Between 2 and 6, the evidence against the candidate model is positive. Between 6 and 10, the evidence against the candidate model is strong. Greater than 10, the evidence is very strong. 10 Burnham and Anderson, Model Selection and Multimodel Inference. 11 Ying Yang, Can the Strengths of AIC and BIC Be Shared? Biometrika 92, no. 4 (December 2005): 937 950. 12 Robert E. Kass and Adrian E. Raftery, Bayes Factors, Journal of the American Statistical Association 90, no. 430 (June 1995): 773 795. The rules of thumb provided here are those modified in a presentation by Joseph E. Cavanaugh, 171:290 Model Selection: Lecture VI: The Bayesian Information Criterion (PowerPoint presentation, The University of Iowa, September 29, 2009).