Gyasi's Blog
No Result
View All Result
  • Home
  • Categories
    • Actuarial Science
    • Data Science
    • Philosophy
    • Economics
  • About
  • Contact
Gyasi's Blog
No Result
View All Result

A Common Subtle Error:

by Gyasi Dapaa
October 24, 2019
in Actuarial Science, Data Science
Reading Time: 4min read
0
A Common Subtle Error:
Share on FacebookShare on TwitterShare on Whatsapp

Using Maximum Likelihood Tests to Choose Between Different Distributions

This article was first published in the summer 2012 edition of Casualty Actuarial Society E-Forum journal: https://www.casact.org/pubs/forum/12sumforum/Dapa.pdf

The Maximum Likelihood Estimation (MLE) is one of the most popular methodologies used to fit a parametric distribution to an observed set of data. MLE’s popularity stems from its desirable asymptotic properties. Maximum Likelihood (ML) estimators are consistent, which means that as the sample size increases the researcher becomes increasingly confident of obtaining an estimate that is sufficiently closer to the true value of the parameter; they are asymptotically normal with the lowest possible variance (achieve the Cramer-Rao lower bound on variance), which makes inference tests relatively easy and statistically more powerful. In addition, they are translation-invariant, which means that all functions of the ML estimates are by default the MLE predictors of the respective functions. For instance, if a pricing analyst computes pure premium from relativities estimated by MLE, the predicted pure premium is also an ML estimate and, hence, satisfies all the desirable aforementioned properties.

Because of the known asymptotic distribution of ML estimators, there are numerous asymptotic tests to help researchers make statistical inferences about their ML estimates: examples include, among others, the Likelihood Ratio Test, the Lagrange Multiplier Test, and the Schwarz Bayesian Criterion (SBC). In general, all of these aforementioned tests are used to determine if the measured signals (the ML estimates) are statistically different from some pre-specified values. For instance, suppose a researcher believes that frequency follows a Poisson distribution with mean λ, and computes the sample mean as the MLE for λ. To test whether or not the measured signal is noise, the researcher may use one of these tests to check whether the ML estimate is statistically different from zero. In addition, the researcher may use one of these tests to check whether the ML estimate(s) is (are) statistically different from some pre-conceived or historic values. However, the aforementioned tests may not be used to make inferences about the functional form of the distribution of the data. In other words, as an example, one may not compare the ML values (as is implicitly done by these tests) to choose between a Poisson and a Negative Binomial distribution. This article argues why such a comparison is incorrect and would be no better than an apple to orange comparison.

A critical assumption underlying MLE is that the researcher knows everything but a finite number of parameters of the specified distribution. (The functional form of a distribution has an infinite dimension). An implication of this is that this estimation technique could only be used after the functional form of the distribution (hence forth, simply referred to as the distribution) has been pre-specified. That is, a researcher needs to first specify whether the data is Poisson, Negative Binomial, Exponential, Lognormal, etc. before she could use the MLE technique to estimate the unknown parameters of the pre-specified distribution. Hence, the reader should easily see that the MLE technique doesn’t have the capability to determine the distribution of an observed data; otherwise, such a pre-specification of distribution would be unnecessary.

There is even a subtle contradiction invoked by comparing ML values obtained under different distribution assumptions as these tests implicitly do. For instance, if we assume data follows a Normal distribution and hence use the sample mean as the MLE of the shape parameter, it is easy to see that the sample mean would no longer be an MLE upon discovery that our data actually follows a Pareto distribution. In other words, since a given data could only follow one distribution and since the ML estimator is valid only when the assumed distribution is right, comparing MLEs obtained under different distributions is self-contradictory!

The reader should note, however, that distributions with different names do not necessarily have different functional forms. For instance, the Exponential and the Gamma distributions have the same functional form but differ only in the value of the shape parameter. (In other words, they differ in a finite number of parameters.) In fact, the Exponential distribution is a special form of the Gamma distribution. Hence, the ML tests are valid and can be used to make inferences about whether or not the shape parameter is one (and hence Exponential). However, when the two distributions are rather distinct in functional form, but not in parameter values (such as the Weibull and Lognormal distributions), the ML tests are invalid!

In light of the above argument, all MLE inference tests such as the Likelihood Ratio Test, the Lagrange Multiplier Test, and the Schwarz Bayesian Criterion (SBC) are not appropriate under different distributions. Unfortunately, many researchers unknowingly misapply these tests to choose between distributions, e.g., Poisson vs. Negative Binomial). Even in much of the exam oriented actuarial literature such as Manuals for Actuarial Exam 4/C, as well as some past exams, have questions that mistakenly ask candidates to use one of these ML tests to make inferences about different distributions. It is also worthy to point out that, under such scenarios, the inference statistics such as the Likelihood Ratio Statistic and the SBC are not only meaningless, but do not even follow a Chi-square distribution (as they traditionally do); hence, using the Chi-square critical regions to accept or reject the null hypothesis is erroneous.

An important question, therefore, is what tests can a researcher use to choose between different distributions. There are numerous statistical tests of distribution fit: Kolmogorov-Smirnov tests and Chi-square Goodness of Fit tests are examples of such tests. These tests tell the modeler whether or not there is good reason to trust the fitted distribution. Unfortunately, each of these tests could accept multiple distributions as good fits. When this happens, the modeler could choose the distribution with the maximum p-value.

Previous Post

A Wrinkle in Target:

Next Post

Ratemaking Reformed:

Related Posts

insurancepremiums
Actuarial Science

The Covid Insurance Refund

May 9, 2020

Since WHO declared COVID-19 a pandemic on March 11, most states in US have been under various forms of stay-at-home...

G2R2: A True R-squared Measure for Linear and Non-Linear Predictive Models
Data Science

G2R2: A True R-squared Measure for Linear and Non-Linear Predictive Models

November 30, 2019

Abstract This paper introduces a framework that allows us to construct an all-purpose true R-squared measure, G2R2. Unlike the other...

Two Notes About the Two Faces of R-Squared
Data Science

Two Notes About the Two Faces of R-Squared

November 25, 2019

R-squared can be estimated in two main ways: Ratio of variance of predictions, , to that of the response variable,...

GodsGutstoScience
Actuarial Science

Ratemaking Reformed:

November 11, 2019

The Future of Actuarial Indications in the Wake of Predictive Analytics This paper won the best prize in the call...

GodsGutstoScience
Actuarial Science

A Wrinkle in Target:

September 3, 2019

Rectifying The Distortionary Effect Of Trends On Predictive Model Loss Target Variables Introduction Creating the target response variable for an...

GodsGutstoScience
Data Science

From Gods, Guts to Science: The Golden Era of Data Analytics In Decision-Making

August 29, 2019

This article was first published on June 27 2016 on my linkedin page: https://www.linkedin.com/pulse/from-gods-guts-science-golden-era-data-analytics-gyasi-dapaa/ Part I: What's the Craze? The...

Next Post
GodsGutstoScience

Ratemaking Reformed:

Two Notes About the Two Faces of R-Squared

Two Notes About the Two Faces of R-Squared

G2R2: A True R-squared Measure for Linear and Non-Linear Predictive Models

G2R2: A True R-squared Measure for Linear and Non-Linear Predictive Models

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Gyasi's Blog

  • Home
  • About Me
  • Contact

© 2021 Gyasi's Blog

No Result
View All Result
  • Home
  • Actuarial Science
  • Data Science
  • Economics
  • Philosophy
  • About Me
  • Contact

© 2021 Gyasi's Blog