A web page summarizing the fitting process and recent findings in this domain.
By Denis Cousineau, (c) 2007. denis –dot– cousineau –at– uottawa–dot–ca.
The
articles
are available in PDF format; the
notebooks
are source code for Mathematica; the
packages
are extensions for Mathematica. All the
submitted articles are draft. Check with the author to see whether they are published.
What is distribution fitting?
Definition:
Given a data set X and an assumed
theoretical distribution function f whose form depends on a set
of unknown parameters, find the best-fitting parameters, i.e. the value
of the parameters which maximizes the resemblance between the theoretical
distribution and the distribution of the data in X.
A number of
different fitting methods exit. One group of methods involves the likelihood
function. The most commonly used is the estimation by maximization of the
likelihood (MLE). The MLE technique is nested in the more general Bayesian
estimation technique. Further, for data sets that have been transformed into
quantiles, there exists an extension to MLE which accept such data.
Another
class of techniques uses the method of moments. One last class technique
reduces the discrepancy between the theoretical distribution and the estimated
distribution of the data using least-square techniques (see Van Zandt, 2000,
PB&R), however, this approach is not recommended.
Definition: Suppose that the observed data corresponds to the sum of two internal stages
(as in a two-step model). Suppose further that you know the theoretical
distributions and the parameters for the two internal stages. Is it possible to
infer the distribution of the observed data? This question simplifies into
finding the convolution of the two theoretical distributions.
Speeding up the fitting of a convolved distribution using approximations (
Article:
Cousineau, 2004).
What is a mixture?
Definition: Suppose that the observed data corresponds to the response time to end a task.
Suppose further that there exist two different strategies to end a task and
that the system can use one of the strategy with probability p or the other with probability 1 – p. Both strategies predict a theoretical
distribution with its parameters. Is it possible to infer the distribution of
the observed data? This question simplifies into finding the mixture of two
distributions.
You can learn more with:
How to get a mixture of two distributions (
Notebook);
Fitting a mixture of (a) the distribution of lapses of attention in certain trials
from (b) the distribution of regular trials (
Notebook);
Definition: The Weibull distribution is now commonly used in cognitive psychology to model
the response times to complete a simple task. This distribution is a function
of three parameters, (the shift parameter),
(the scale parameter) and
(the shape parameter).
These three parameters are distinct, each affecting only one aspect of the
theoretical distribution. However, the MLE techniques (all the cluster of
techniques) are biased in the sense that the estimated parameters are
systematically wrong. The exact amount of bias is unknown so that the estimates
cannot be corrected (unlike the sigma parameter of a normal distribution
which can be unbiased by dividing by n – 1) although we know that the bias
becomes vanishingly small as the data set increases.